var/home/core/zuul-output/0000755000175000017500000000000015155373454014541 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015155400336015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000313366115155400260020261 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ )?KEڤ펯_ˎ6_o#oVݏKf핷ox[o8W5!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kf>|\@E1%]˜(O)X(6I;Ff"mcI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|iZ~hal\t2Hgb*t--ߗ|Hp(-J C?>:zR{܃ lM6_Oފ?O1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾg}\Sj#3hEEH*Nf äE@O0~y[쾋t=iYhșC 5ܩa!ǛfGtzz*з 55E9Fa?Zk80ݞN|:AОNo;Ⱦzu\0Ac/T%;m ~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0,d%1}Uhs;H?)M"뛲@.Cs*H _0:P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6,֗&̴+s~x?53!}Z[F)RH?uvͪ _5l *7h?cF_]CNnW)F5d,0SSNK9ް4:ÒozsB<^+鄌4:B%cXhK I}!5 YM%o<>"ہ)Za@Ι}YJz{ɛr|hxY/O$Zøu32EʉD'MS1}t i:Y`cФIX0$lη˽`!i:ګPSPٔ3@5;ȕ}PkڪH9' |":", 1Ҫ8 %lg&:2JC!Mjܽ#`PJWP4Q2:IGӸۡshN+60#:mufe߿~Y,iǑ wVq*T+ w%fx6 %u̩1hӰc%AYW ZY~a_6_yWf`rVA,f=A}h&VOK(BSsǽҰ%>kh5nIYk'LVc(a<1mCޢmp.֣?5t罦X[nMcow&|||x:k/.EoV%#?%W۱`3fs䓯ҴgqmubIfp$HhtLzܝ6rߐǽ0 2mK:ȔsGdurWMF*֢v|EC#{usSMiI S/jﴍ8wPVC P2EU:F4!ʢlQHZ9E CBU)Y(S8)c yO[E}Lc&ld\{ELO3芷AgX*;RgXGdCgX JgX2*Ъ3:O7ǭ3ږA :}d,ZByXϯ&Ksg3["66hŢFD&iQCFd4%h= z{tKmdߟ9i {A.:Mw~^`X\u6|6rcIF3b9O:j 2IN…D% YCUI}~;XI썋Fqil><UKkZ{iqi :íy˧FR1u)X9 f΁U ~5batx|ELU:T'Tស[G*ݧ ؽZK̡O6rLmȰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|miȄ{g6R/wD_tՄ.F+HP'AE; J j"b~|!p+,ICE^fu `|M3J#BQȌ6DNnCˣ"F$/Qx%m&FK_7P|٢?I-RiAKoQrMI>QQ!'7h,sF\jzP\7:Q\)#s{p'ɂN$r;fVkv߸>6!<̅:xn<# -BȢ1I~ŋ-*|`В~_>ۅm}67X9z=Oa Am]fnޤ{"hd߃Ԉ|tLD3 7'yOc& LFs%B!sRE2K0p\0͙npV)̍F$X8a-bp)5,] Bo|ؖA]Y`-jyL'8>JJ{>źuMp(jL!M7uTźmr(Uxbbqe5rZ HҘ3ڴ(|e@ew>w3C=9k-{p>րd^T@eFZ#WWwYzK uK r؛6V L)auS6=`#(TO֙`mn Lv%7mSU@n_Vۀl9BIcSxlT![`[klzFض˪.l >7l@ΖLl gEj gWUDnr7AG;lU6ieabp៚U|,}S@t1:X _ .xI_7ve Z@7IX/C7@u BGڔE7M/k $q^hڧ};naU%~X!^C5Aw͢.@d!@dU}b? -ʏw |VvlK۴ymkiK% 0OFjT_kPW1mk%?\@R>XCl}b ,8; :.b9m]XaINE`!6uOhUuta^xN@˭d- T5 $4ذ:[a>֋&"_ }Oõϸ~rj uw\h~M il[ 2pCaOok.X0C?~[:^Pr򣏷y@/ڠ --i!M5mjozEƨ||Yt,=d#uЇ  l]չoݴmqV".lCqBѷ /![auPmpnEjus]2{2#b'$?T3{k>h+@]*pp桸]%nĴFԨlu |VXnq#r:kg_Q1,MNi˰ 7#`VCpᇽmpM+tWuk0 q /} 5 ¶]fXEj@5JcU_b@JS`wYmJ gEk2'0/> unKs^C6B WEt7M'#|kf1:X l]ABC {kanW{ 6 g`_w\|8Fjȡstuf%Plx3E#zmxfU S^ 3_`wRY}@ŹBz²?mК/mm}m"Gy4dl\)cb<>O0BďJrDd\TDFMEr~q#i}$y3.*j) qQa% |`bEۈ8S 95JͩA3SX~߃ʟ~㍖›f!OI1R~-6͘!?/Vvot4~6I@GNݖ-m[d<-l9fbn,'eO2sٟ+AWzw A<4 }w"*mj8{ P&Y#ErwHhL2cPr Wҭюky7aXt?2 'so fnHXx1o@0TmBLi0lhѦ* _9[3L`I,|J @xS}NEij]Qexx*lJF#+L@-ՑQz֬]")JC])"K{v8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGP7^s08p15w q o(uLYQB_dWoc0a#K1P,8]P)\wEZ(VҠQBT^e^0F;)CtT+{`Bh"% !.bBQPnT4ƈRa[F=3}+BVE~8R{3,>0|:,5j358W]>!Q1"6oT[ҟ^T;725Xa+wqlR)<#!9!籈K*:!@NI^S"H=ofLx _lp ꖚӜ3C 4dM @x>ۙZh _uoֺip&1ڙʪ4\RF_04H8@>fXmpLJ5jRS}_D U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0NH;ݹ/n -3!: _Jq#Bh^4p|-G7|ڸ=Bx)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==q?{WƱPz;| \;_D[T/BI GH8@"t*"9z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W>sngir^$W v:?_ͬ5kݰw[!$s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O wmJeATj'뺱dvj5I0|E,YI+u'>F@$;JbyR%2dC䢙+kSW#lg7k]YsEzeIW7&s?RQw`A*S낎K)> Oa&wb"*PEw(bֻEڞQS ғ8)PT9l9@KZ GY~MZmŊ=p]}z1'e;v+]P\+j' R]5[E2mI[|e'!޸E=o!C~/%y K8|k3lEPD6KBw\ kai2@)9Ś%xZώg"fE$$N^PD՛ۗ|, ,VKG6q1~/ 'ߟ|g9h|v?t`J!Vӻ~=sRӨ->p \W*Ԇom6<IrmD.JrD=Q<]}R XȜvu}kWň!#&+DcVKllGdEްV%Ff{!aE[(P3􌠼)C`}z9BεV>i㮼|J %ňLʧ0q_~{;2 E(4jTH}-"+ ($ T4o3T<<&quQĢH|u#kppH,ҵeG(I!ғȍk̤:x؏sl :($ڟ}^G )W!1!"9Ul/< )[tc'R)G= K8Ҭ޳#Ȇ"E_؛j&2r+N "f/ɧCƕ -%0'2jD.̠&gݎIwZĥ})! in_[Q8RL/r[juF$Ll# T=Pƌ=qu6j~FH発;~:VX0rn<А5HO 1A! Fh P𠃂2.>{7i}ħ/^ @]t̶7@ NjOK5/%RTgc: SdXMq 70UҽˎL1oSpoVf0mS|ԆV.j%bZdl7U/m$\bAb52NqH〩8IdQbu tm X2`c a: jCLm!݅C4 ̘ɦWB%`Cs K.TnZeI6I4ō4Y[1$ytG,#8p.xKvbqu@ao@,5L2_:ܺ.h]dccU˼1GlM`lkٛOCcLgF=SL*pF Sy jՂq=ߥq>42S&g_""m=^$CH]'Pka2ϣ]Yk4 ܔelťdH 󬗀$S[ۗ+<ƴ6n+ {<!zP׃q&n '|]Rk\)-nEl8 ^fc#..Pmֽ\Q=Q7g]JYl.߽f;V><+R hjHBEd>>ku'j)S,X+Bzqț'R8G9걡Ģb@E{K j#H([f#5w{ 8` -ZT pu,D&{g=c8~R#1pآ]Jt;Vbj[cxcb9cmjAۻ+ P/J;l['eP"uH7a՞o]F;%)"^В`MGW%(efT!h}ha0L"aH j9tnamEw^6ABy,(2dޮ}.Q`!`DԤ]Wř⮴xz'g򢣉d;hsKM-Q"Z4LȬ  h J}{]OCUzZ"=8mA>F]Lwlrb,K/˸RRїwm8"łHի ^Ȧ`a- DʬXb (pѤ;t+Gwo3[ݗ1Ã{0fAϷ98%TEB/ AB %G˴$^8N_Z-~uuu=LB";f?4+}򬹱mٛqk Ow'_6>ػޘ]m}麻ҰSp;?NFLg}1ݹ戌W֕Ύ%cb W {7#ݕyA\:XC^f!nyA"MB\oV,hEf=w(2vwgVC3~]dY+ 5Qܕ~R#,E\>h3Fx&%ǘ:dGz3LoZA19;,1zّLY?[F^:2vs-"e h)D3q) ݱv{[{Q=CtUS 'ey5Ha6&i1h_fsZ0";܁.>f[r}iqh逻)e.lףv(2K'sK:HũZܠ zOٻmeUq&f $@D]Gdswl6zn[I Cy 6[K.t<)ax&߾5H@Ӓ*b]*?w<Ew<Qsh f@Bf*X'F8L>ygepvh#> :Ez燯bԕSW ͡fPͬr 7F| xipK^Mg bh!# n\< چbB5t~:>/^Gm`t±Rzg"{Ο5RhO "NJ $[A@D\FȁKOZ@.,*hf}Jv@ "iZ0Ӊ/zH`a vb}0eVmr \\ltػRnŞ$B:R)0KQ 7Hh3mm^l c|= o0|b-*|@Cv%Jĥg&)dK1IbAX` 3X`"m+|DǂB\9 *'ei;$]9 y7BXhښS+40 B'dg%BIʓZ@F?EZ( Vqrj , ]q;yũu#SH"&փ9 nbጺHay]̦y1[WIEAssF%s8]Iw&2s^a8^sv87l7s7Xz5J97Pڮyh?GUNno2qW/Ub'qCCx:0S*5)NA=L5v`8=j^:ڎ8op8XxaἮHVUG$8+-%dxpW8T=V3 ̾Nrpʽ=F.h@E1 0̑I`χrC4bAY`!AF|dZJ&/D#9 mcfo2P ӄO ddU-֦f%o~gZ}>*j0԰z+LXY&c3uխ fd`|lcVR E/n3"j@+ۀRt45kkQI/ 2€jR]%1_%D,ؐcudC'1){rrL^p5#pHh";΋{CP i2XкѸTSA06Ү{MjH:%KT i ZHE9EYOo`#؜j4A8ےOt:@%:@g@sǖ.U'+4;3݇ch5BEL@Ũ}yΩGUoaG-I;*fnZd: ؁]tSrH% (zjm0iw!JW->sl_bPT|(IknK$GV︶v &QǏ@GqK/彡{=A:L~4H];J{QڣgsE>PW/}pf !C6.\GL:qg&{_G,pJ*%-jBljeQrshTQMpFƌA,:j\5d7d_ff@YkiZk}Jx~`ZS[+:[G-7V-X6>;Ǜ/1L.Mfuv6oSL NK܃w9o32҃qFl^'YB. @a崀iհ85{^:B K^³gGØ,K7L0i0Nhy`!Mg;_dY3";`)>%?Ƶ}+\[qݠjpZ/ 9|7{BuB{>MXg)<w,£f -" kA|)e%ROD)L "֥nNp7ckEɷa|^t-,%S2a@a>DAmAC蛅woE^XՉ VЯy ~-C i<-=%w+ab1# |^5R/-a-k.c.5^ywoF3 -0xrFl s.2Le/ Z8qg-)o,QWl #ߌ\BCMsVngbQ Ӗ` }$y Y^\d8o;X K*3F@}!wp0@% ‚Uq< I =)ؤ@@wދ*pu3qR=g"`#vZv:?H+ wң64fL9A94sN\Zjk׈]WK3YXϚsԦs=ف57gT 'I _޲cJ'LDs]ܤx2gir/ӛM'QCmM;"DeIbtdݱa6Edtw LSᗉ#Ĝ9T P̋}Y"bjMᯌ;<J蚩ptxdMP^zpSP4&j%H)W!?4hi[,t^/S1Xƒ50'sA Sk6c"osJxbה6Er´Ĺ ZHtd%U+5ŒGݞӯnv,_4N uNkfB5Z'HjGZIW7;P:&BEP,8gRTɰz[h 1%݉LuR?߉č\#ܗמ| m u쇞QB RU٤4BSm{A*FE&2: mO^'unoq}qW)&㙡 [Qb%`-.0|O -BT% V:j- -kåZۢdFssnȳ LCg@r,Î, .~?n>dT&thqB[SCq|dE*2,#5zwmԀᬪe} p6e;`$u-CT3̑p OfCb4qJʣg5~x|i îeH%'QgȚh]=/"nVOEL@m߱М1? cmȻ޶W$*䵘2-␲dk0}/I8tHl3{ާ1L]Npгl7]S{YVX}7}Йd~Fz ۧYJznX0q~MX_3~^-}=WO)9>K-#,V}Kܼc{gs7?wyoD/a@ A<4I'K2E/P%蓃wJ$$HS`K_ڦ9(C/ gW-2M)`sNFεPhMGمxʹ8" h=|;I3YȁaA_k֪'YkY?uQ-1 FbA.A*blQ:s#nDh ɺ@'BrY4`Xe+ZPhAmۼKQSjyQuեILJk,l5(38H3:IM I'c>.^\x?׋`K{{QeてUvqd \t?1$)/abS1Lͫ*WLBU`oi@TRz"òC+3,  Vr\Sxđ=Ǘ.8kY*ErkR;WcYK%0X ck^U^=Ky9fA7G0{QajhXoP &erQ >kpc҆կ稏G" g~\/C*K c)[=olGZL9b"r';aЕSc kA5U֔ fb㪼v o#9qwp W[ *@#W\!P$v<%&+~YnCqPbM^ԤS`%cka]{iyyQsbso\iN3Z>~BkJ~#G6.W<n6FX\r.91SN^;5.n:7|ƙ(-jr O%>@n; dFHay0\k SX,C\t' :3ˋ{z>'"_]=/#f`Fs>f-prׅIgJހpS.E! `F^d4)N7ڥ%sRsFeqs+@d%!EtcilpvKڠRalX\hVN-ά`|jvX3pW%e)N \?.8V<4WT7>[6[s ap;-ӎf,g;:wTbƱG8Yj#Wv9J]nޙ1,x6.͘Jk'tX?,U9@>llVy6E:1騈ޤŦp4'w9=o}p~>=Eo@"m:rU t ftKp/ǥX3xI;sY=8znC̡ ema;4Kg]A ]JyN懠x (NT0V{쎰ڷ a`[0¹^cAl1k0H JmkY#sx991Muݘ5r,v^ۘir1 #FOYMM3y-yXિmo?5__Z.:U/(ox uYLjA߸#җSW-_Cl[`H(MAH<؁&g[q1KӃV9zp+lAGGJ@%S:ML3`*4蚩Z7g~)ag_. /&h/P]piQ! c Yz5 :=20eU %3q'>_ DL)29 x2_ÂN}wW:1S)?mB0EoEy>aߘ C3rkwMga@6EIX|LDt%Yf.ɻ]7 yRuO,757w|no0 M!DX\RFcݤ--2n*N1c]{y3kB\}MQ$QT\[Ρ9).Sֻz9&bJNAsSzn?X{]D73שO`z. ^߅Y1u d.E~>oٵf0 ?1S` ^G'E'oU=X&r_>2,^Ơ0D޼-\ANh@WD]*3>Nx<|c|SMF'IW}'.Fw||lH2tW ?C9sSp*ٰ_NL8'fFq Wjc,LT J+?c6ˊMSzRWx-|/K-׏W׃&Lb쯕UCq}8U ˿8dXЯག< Nl.!3˱(d|4}}ɻlR*KN/H3PZ>J(ݞ?C}ʵOgS`M==M3k^ס | 2}BZdt;:0a2(8G܅d8*}pL(ͣRK(7՗_!H& A)&eq_W)rOB!قTU.s*ҢZ"2S|RtI`t[!KDb,Q{ûiJki1>G:@K 5Sи'H/]TwJGpt0;C9N>-$e }>dkR={G)d=k¿Le>ޝƗ5@X!<=̑Y"V;G'59ovL|m:H,'@v-5{oDaU:8jrD5>,kA>0Hzdzx2Tb8Ϛ<\bT3!(aA T.h8I4Y^0i~ӘDD`-^Xm =yWSҲַ;|:K!b$:dt&7 aܻ󰸲Ԉ32 J1ua77SkYJI-Ȭõ0|@\tl 0-ҕGlBVbimrL|il%-[(AQ_ <WsMNh%N:O]TD2ʶ_p3Jl p2$`dKn2[e*NM{k _e_#ՀtKLPX<@r35_۟t ،H_[_vBS ꜯrYhq4uj'y-4ƍE *u{>īڥ ^b"mbܿBo4ri_AjCAx3/x߀M&}#%F{)m*q+ڞɉ\ޚZzLzKb^Ѣu(nh%JJbsA)\<2J#UK l] Cb,@ͪAZ>mA8EYV5,{,؝K3k,v qA t/p%bMu(X% ,=?}jӎ TqNOzqgyK]}Ο, "DL[-iBH+t#K4Sh-N(ւkk]Y2vH2vtDnٻXQ _YdT;Cp:YA޼- @Yg|k`v'J@0ϗhUFJ6(P-[_K@"T5I/of4ZLhv\jLJ D-v+̓ۆ҂ts5ٚt?)w[hܱNR$_}-0L5;;bس/HRpn: &tț·9[͗b|YGP+a2H.6R$j.EAn.EVإ !.^1:RY◻KdZUt$19bᲖA9)6Vx\h0kq:o亶\M:WR~L6ʯR*E,7w`vA p!V\idѩZeU]Xk+ũBFzn=`rQٗ^dgZ$" xXZ{湰 Bަ&9x'$7IMg(Y<N"guT1\`FĞ]:P+mK7uhS\Y1`C!QDPqPN#.(`m̉V (CJ)!-wjCw1,_hA+͸K%\N%;O~=V6XE;1BQk4|4BF!!,xMA q{4JܪA)_8k0t(t{-?&{ꣳAdz1znN9T` hc?."J! G"J +Gb PđHPIB M "ChaP~nR)l6;F8(.9 t"r̈!*-kJ(g>1O ?{Y\ޚmA-a߆o;Y3 (;Y9ʧqZ (GDrka4]~EYA;g)Į52fW.AԄq]R$ȉ+0Ɔ Ar%eJK-ǎYȵԈ8dNzSDI.f{u"ע?Pg=xB6 wP"ȜyR@8f Fc7B # %1ӜSA(\@h/x cRh"7!빷@-7g\#VԈ-(#"т"/4 ?{WƑ ! wϻsI6p(0kI4vyI  J6wYhsd`?ܚNGBpޡ( vg?m+8z%0$#弐eW9ItoψlNnm( ѡ Ia+bRJ2uFwFiL(lq14:n 7ov|m;\ K-JBzQ .пc*Bğ+WY|;U\I39! ͬ5+UID$'O'v}? \ mZZct`jpݎeV *l] H{W)KK2ˡ*5 '' +YP@o4sDEuvLV߫C{:WLm@tH JGCk;f4S] +ڠ6Ro Q|?qqDm4C |rtj6Yh{ FWg7cz}Ȩt©* 1TfUe)+Br*}!Sf._³7,U~< m䘫 ֖yi9C݄28&6/p@kb5c3U*cEn g@x %e+yJ(. dq! 3+**N93~IΒ;542u^ J|IB6hcp)KkY1V<-yFֱƑͦ ѡʥe,XHSwgJsUyJrawaWƫ(@ u&q!iOZmK]`rPx=`ڋ53]XZ[ڠUAθ= +&X-\LX83S3JEx(\"[ggL; XmjC`a6p6f__ƭwt~uyqu9Ol#m}!o&W8 k:ޗo<{Y?o&EoL:eTҏ;y_"0DO̠T~36 ~b뻪u\3~[-(JTD=ڏNPLSۜd>7Vr4}b<д`SgCARy`j)&"BUܭc~MbZlA&"T$ZhK՞SkHIlRS:U!SA"_mJ2EV[llgVU3ֶu4잃Ie:R3펔Dډ g0?c1jDW$>L3`VTPF$N{KHpBX\Dmֲ .H#yp\34ܸ ; n.n;/lx4z톮5_BYT0PQRޔlNlB0`]s᥍MunDi K_MW4C1- hwB{'xLjd5<=_o| DU\3Wr6WB>x?{o_CNS33P{y3ok]8G6cGx꽽ǵCCa;WNǃgM<_i̿C?h4oxGk7<]yQKh=Z,ݨnU+7wٜ53 m,{0O2J|?u+5.sR{?[7QoMCn(+b3`%^e/Ը^0Y.ѣm v;+'\KRTbN0i_Ms3|`POn½Iv)ç?7}p@ƝQ܉/O?}|R}y2HࠏMP wx8<]Q>d'%~j4 O X=&@ۿdUHxP!AO~%vEtz eEt+Ŷ*VMmYІhB}ʋM}ijTbM/f)gҤFE!|$o4Tu/ɧL }y|V C ~ngww \;A;Oww&фԧ{|?~Ai󳏧W1כ:|:8Aߣ|z~Oa,;CUd\})tzi3|Y2EjvK3ehQZk"IJ<<djtox0< Y3z5:=;>N\ZxF=*}D_fC4\0irjS%XW9B9Ӂfp!W~ Ny9㓿N?.~}ϻ>|k?^ Y?-0x=CĶZ[[. w_φ qOݰbA017/\\?eXO^^ῇOFf4 uϧ[AL-]q11ޘaZ2Ǵjl yy~a7iv[#wmB:w'#T.5Nm Bڴ^ Z_ԩJwz,u۬9"΂l띅8gAМxy(nOY Z0@igzrUny)g WFUVLXQ)"vI>=ɷK`#XWe<,^i_8f)mbs#ELg~RNU$=j;:'F)ƚUqZ0ptNmâis*ii<вQLsN)6HGN4z-߭G?v|G]ζ$ZTQ&GQ46uEX(ԒKme &ʷ9G.5V hf`"$97#=کep.^_`7+9#jl<Q4Qk73{vg 󣝍m UEfM}h`is^\x4XїöUQu6/lzT֨.WY4`'Z^~sP2LƓ8" Za*ҥ]Hp8ҥ:mKAm1zH7~{Wtլv$W> ZGnzQہr!e"SM=0\8 iF״Q6bkv/ݍ17w*!2wqs sdl %u%5*ВCxg\2 Bg<:2جQFg<U4 |7 :F7{864Dl<#imaFt;dvI,he%r(05*LrxwhpUB9 ZWHIx)Sd#jC&(4%P1ANȳsإ^ i9ZX4ҟ^j;5mڟ:]J<^BhxY6K6L]:d3[G++[vH 6B"U W~hY 0gpoqmD˝+(vYL+"2S='u!9Q9Y6K˴hICJ~ K wf pub;q i)/} HKǯ06~f))~fmll\fsRV"ytc䎾YjV܁CέY~H-m+&!)/j"!>4˃Fj#y(R^4yp)yq[J=99A!yH"`E$M#x`cVn, '=Q L&J*#B]1"wΗKG ju'>l=IO|MIy Z_zUx$AMG'q(05;;ǡMLNDtqm͗pa9 !}x,- qJn_Cq&~OuiмyAՑAêV{Eճii #fY\ΊPeU2&ɴ<>3,+Jh-]qْEoyܙFS j'Yb F !8kOVK]iǿ><BGmuOuiyhΘfOYthZoKx4Y.zR6H:.9fl6 j(0-[|OxZ;"\ѧ&S_RG_F:5F/o!;q>@G @K sݩP6Hސk֤uW멗#i2S<'RkN3> aS! AWa-LT& ֌*Py<؉d?Fi0GvP)tǖ,6yCi4U KJc8sȱRrAaN;H0Al \ˀ= ՍcQ[ iGN;"+skn#W]Z\ּo%5/~*+NT{Fj4&ES;5xMZܼnżoo^Dz-Խu`\Ω{w&95N</hO J1?Q02eY,ޜ8Pw|ϳV$_ W@BxW|z@ӝ2}z9zV7Hxry&ڪ<D;{:zu{Ĭ%Ǐm\~\zm?M qR qOĥd8{{4a=[f 23h q;ādK+l[= eO#nY^Gk{Լ5ҵLe{;]w%]ٰ')% s+qR03eAHls4eYI֘Sg^0i˂3/eW0~΁t7`3Kނ$X?|RL+{1 +0xuXY9kz@nC=픂Fynj^5$6IAl"6GzRmӃ% g\5ly|x.Wͺڸo]lăVUdPوn7ab)&S+N4o.юmqķ@Կ ٳӼf]&ӟ2xż#'}̺Zpto+8emvG0|_wy-iBP u='ñ$} =rg<_v[Xb'z=z-omݘ(kk7 W}ԍ*X܄B9||?l z>~xn>~xx/O} ZZ$KZ$ YK_ Toz%G'1_\TIxHR$9m$I[iq!C$7#TAn}j9jF,1i'd#GvȑSo=-$FQyHㄑ|(l/LMeIjF]S}_cNӵ/zǗ6>X+]߫{X%P#v $qԘ`$5!Ը@j?̕j/S *KPo4`aۈ8gt2 R]pg!lSe-mU)zam`@3i¸HIN"Ѷ+noq{ A**V܏3*ˁ;[c &}Dĩ#ST.noq{eFE=PN" _Yԃ'2R* c.noq{2*S o,noq{hD#!-nٺGE OJW*~,^ގɔU3g ʨH㾒D#-^z#WP=>󗪚,^dcdJ@ɔ́;u:۷ knn?߾(o_9i/?yug?U?IJr߭(nk\sϤTAsAu>ȏV]g3mh/B J*|Gm缲uToմ/IOSCĸcРGH#%Tq>oq_}K몹r?o6i%:mi7[;.lZ[[2oۊubV:ҮV,zCW4}]n%@GuJBUPczxT/C0)\MWX÷C&k} z߯Gi_sDn[`( Mr$H6q@J|6Fɕp4-_A%:X>7)#B<ַbi-0fZիO,#ѡ!Jl#3Ktt=DG.@,CqSp獵I4!Hɗ $!kRwv'AoxD"΀IlmĒV|;Ȏ@KZqٴ|ש'J+zg.aFB䠎ҊF!‘ @[-ٟ#ͧo rv Xk[ wF*V`QmZӷ]+ !D*9(ǴKx!ò!T$I#YRy6RmKm$B/;"n`8-&4 !RDBFOcڍVa( #HriӔ([q%lNs,@kq7p,i$Ɇ%lWsaɛ,lUX,K}Fb&dH>,!’)4-7+,9xu~nkkwk;s9XT>J_5rWZH]p_UzZjG/ 8g!7Q2>rF8cJ6k zo C7;<_8em*M%H>|/oj@/](0@Y>H6|Qk _&v誒=^/]s푽h-*qCPT9N%ɇvڽIJBBnmqbёZFW #Fʾ h^@7::oU''1qL oE؎bUkJ _}{uPCC!? R#@FS5Cx$?C@,nJ[H%F,qolm]Vek0kזQ# `2a&Z'Ɇ\e6=B0S/Vcl 1GآJ(*3w*x@<^w(nV W݄/B!?&!H6pgLs/M4d>f`3 4$I24̉WJ!v.e+!!A^ "3cH6@+u Xe[No(׿ަ F mO%~e ,$FUcRK~| aw{g@+Z6U!4 TaioYRcjWV RH:zIX#HScLV9%;6#9\:$9F:De#vD"$vLɇr+&QRTLe~V %E@$I#\Y -I{g=f=B̆C c@0e$Pxup>Z @$_=E{qL]},>9V؋I)*C >6}2;~JC8f~H8ԉI05i$L:Tɿ_nǒ=lunSA٭z xƔMlXlAF&yffnAÂ46'HM0 rDrx@sUzs? r1QŒI C+aw!Cc 2EƁf"QP 57q$U{dpo}ߙ"[YާC2Hr8c㖏$@L[(Dkk{K4᳋Ϩ,$~K̒ĭI#pKcέ><݅[pkG77F-1iAs8I @Ա)utNZ2 sfח1~%2d>4$L", @GKx!x![Xâl Agb4# DŖ6Mt!pҪ2!&Y%.\&)#*F#e% LH3QQn2*ۆgKȴL4yH!{KC#Zv)6D<(g wK`\햩%@Ng miТP$Ħo"З֕bv:@ {eQRǙ&MLg]mX?gaEI4 x{2+^',=usG9x K$Б "`$0Y[ܡ5&/\ܡCC[6M- g *Ď|J6H6p֔<&(sFP U.0dS~^ ( _lҷW}Ză6?٠K6ȒD`$\;E "O|Gã;'Ô@lDNx]yT7o~*-$}^+/fb#xbp!Qb%dJ%ᡕ'Gr)Bru >>|d#dt(`) D@(rx2e#ËMX#d'iV1n)X5`#arvSE$DdyWA)i͡,`'Wy(@t{C`5='qeH>`!AXxl}qw%郇 2 : >t?H6&x9_kD"qI'6,)iAB,zD>rּZ"疼}+W ׎ظ:5뽠^T G#|a CIT4W؛M)P]\peWn{h b&%Ltl\A7PK.A'/&hA|uAOGA^)~xqC$L)zNPdT!%9}8݄zx#HYMs-}e 3RK0ɘvۿ<5=kl fylu0gw2gxak}qY>iludkmnVe]*]u_T6J)qMq)Io3$>44i4 @*Nb.h?-~ i2ހ-cOd2z.֤`=0\voc0ճ@x<_8{5ax:,n|bMj𽻹ivI3MW>&&E YB_4X#:(?@]{!Cb=zU D'7 O ęDͩTș2mbCؐ}ɢ@B )PF\N^YĔ n23t l8 8ALۄj6&(">k ^0tͣ=ߧRx0&%al;bu#OTB@[9caM.. aاX({"cT,6t5Ya&˧"{҆6)RlKbe?VbKQp_fzPrsF6<,%EM`9m+6DcYSdܤz$(- f7c  -5/>!XN.UQXO&]XWr&_Ec ZdM"ֲ!kn T9HJzH%46s#Yu~GI_Rؗȏ4JhN02?nzt]' w%&U*ʫD0^Քկ2EZ )b(C0`,<"\K)B80FLcE&Z32>8:{xH`,铁g#yXqm7+ MAc)̟k\SeL?t4Fhrm+P/%GM'E32tJ^'q11#v&0;&OÀݺڡQ]7R<K:l62S>^0[Npej*_?U˶AP򠗄0?$lI1͖Ytퟰ~`3| j\vF+$;i~k 'Tט&,&X*R?pQI`qYEZJKV+?`7|hܸYZ{nƠh^Wvڗ9 aM\]sҲ=ZR}qTܖ f1sf0㸈qV qլ4;pA۬۶RzMεiyjV5(nݤIZBNu  VUoJ9hxXu#bRs͡9f~ KB; 9U}.bF`/4*T|!J3 TGVQͬ89fṊ|MAeG1z3 T~:3b߹ 74&?U8`'aS/OOU xY Yᘍl|'P8@A i/ 5ѸtP?l2<2߉̜Vz,J7Eya^Q1sj!dQ:ĆK,o.X^qw>2cr^]fno:fFtl ޫwoElbbxvn8vn}>0ՄXW^q$S}$a6(jkY;TЄFGmz, :AmJifmGI $ c ,*o~Z,x|BJ[^[Y$C'_>/*V"޷Sf2 9,Y,õħCxD{28L ɻj%<tcčtr{,{?ܘ=a6wO(83uf2 g)li)pɤ[J\I+ $CD/$(O)'>/hd%(?7Z^]]0ӾݳWa21uaK^oYo+4'0d|J;/Ʀ V\_UCs7TP;sS,ӡօ9_7{g{̫WKgP ɺ3? Կ+W,``vwpԾg;0; O8D,F:F b #TDQL@Xa޾3a/:<; {Siߝ{񷉽<m\ ta!ihM"5 fLf4aH- JLӡay</8g>$J 6o1Wo^>Sn٪1vywŚ*>>_?ZK7g_~hWwlh4Nl #,CRMyu|f=/jMkr,{?VnO7)}vFm?5>?2jFYLJǠɩ[Yv\!2$7/Y-/Y3Uvmn4,6'ĝl.n_o}<̇M~'t0RDReSp䌗xQvf- ۇAl歇5n'&~Оi$N' or]hQdtTG-2՘јZ>mغ2=<bYR^T瞫Z\Ł䪣j4r168FI0OA/ ǘ[<Ɓdx?sܶEk|&0(⤣10Q{06_ ;O8|7ax~$ѓؔ׮(+bVdZ2 ,SŴr39=17g`]޾asMZ2W*9V$Dt!H A 0 |^]L?odz*R׎jJ7~qe(&{cpXrKGcy+u#iIq[o( H!T=k1ߵXZw~ysJص+?A$( ӦHiͺȳN:#>5/pCs(%LKrX|A|6_PڎA8|}l.|8~{Sz7cvI :/!ޜ;.,=q}bq|K1/Ͽ_ ] #r@1J8.,(ueWˢJk%LƳZ:#>X;?d(^Nͯc#[kZQШIKB B:ţ-rWb {ͯQlqdUfvh &Ԑ)4\aOc*IuuS[P&{1dydT$ '.ΛsGcr qCY.Ɠٻ>jeO`n[ m.oAb9cL%@/ێZP:=C:j$'fK3y4 HqFhg@3>q~lri5d|Fw颐,1"'V71">_G}@ƀ۝Yz4d}8掜ѪzL,=|]>9HaI&"K!D:hVVuI*ţ!1E$OGR"hJ˜(g^G-2U7Ts !7Itwsc(.wzR^roaثnpa Kxdfg@s| u^\52IIPܨ3 x$ ;ZyDQ;6T;Lz܉Ntz'0%./%Qt[K#TxlQqH784ʓm^yRNWiڭ;G%רנF͏z"; ̷Wëz50^Uv$M,YkNJhJȈ/s+I:h/G^VmQiWmLp dz".>lOwrg#.ӆNx}65p{ 0  [{ mG-fdss͘xL$2F'ViltEj"@)7t8QF}a r-nrۊI۹QMnH7g\T1I;b,QWc/zx*?Sc@+$,wgc8"?$">n'Ȣ2y|-ѪzaT#0V5.]T#kD.7MW?<ȼV k_]Gtxё໽P<- 3U"hnKOD~Nw-j!ww8)J3<gl5q'L:m\޻ Ԟ]NS*ERLvEm<ۧ!hk`ld8ƦRN^v]m&a3QKj2=alp.] ϮX pq9a>zzr2E 3KyEmIVM^HzB]6o"=^5g B|' XK9.RuG<#ŗ;j]Mt=ᄄ_GQq${-hȥ/`~ϒq'Ehד|^9NDߘ~ͧNe*X[ (8߯ B+msώơt]8ݨNwonyt=l0XsVC3Ƹd0Q{ | S33.$\N$j8+͘%"~.Ho%W8kܨea˦9*.W\l܄->=舉z;hD1yx3[l9rEn&?%/I*," 27 PClRX=}b.Gb`M!AHXpXq<ۈ;{cpXr68ƕ.[ҷeVUh06(V_{rIJg KMl1Dऍ2wسr@s6xJN,egKP=F;; kL)i'V ^55`w8$M#pQCYn:?}+WLVT21@\Lu%hҥ滶ZGR^S\ u!UaԔlp\X%S"RĢJk% X1 PZa*L9awߪ 9Y+?̯%PV4*|DOS$Zd D?`r8F)z- p^ձWO_f5jෙO*/ONL\P1 I|z<35M2bZ$?/8m>UU6iQ6T_;X.Fl$"a9fk~8%t($"ħm>M'.~__z<'q9b!j桁gYͯ_e a9\1=ffz׬^WTQO/-` '`"YDKi4j}" 0rS8^=oy{zי<[5"XBݬ/LQt`䵣 1 XbTK 2/ak(|5f(Y3"(S=T)zƺŘۣ+ ?%ċ,p7NaΨ/i=(ˈ2 | k\kRRcbl9}H,<+P" 1ERuAxLiњ<^S8mKEv%VXH6tPx1AF8{(=2cPiv;IiM]~0RDReS -=Mx^ǯzM`rwdzj__ҭiA▊ҌXWYz.G'+o$O^HtFUuШ M3[ VouШY#6f*OC2~nbMa!TXb(h RoyT#pO5 | %XTBZJMIû` >Uъ(hl?[ eZw*ѩri(ڊCM'YЖy2K}v1;0[cz+naHsn/ Hu;IEem]n)%`.,UXކ]o>5NSgOgIjr,{! An3YdD ghBnُUOyCf1w^xb8G)y4`ƷM$2d.M=?ZFD^O0gXՠ3dInpjUi;c]r}O-  O ;|p+j7wȦyؙ:sYsX4ȣ} tw[T*뫧/ys_u׆7SbEvF_gy„BB&›[ YN91&|co[x K\* *V3d' 4hF2NgY\41 'K$ [N; Z*<\Bg}$|2+ -&B0ve-֭u~8`Ō/xȢK1Iafb[Bw^Kb$Ip}~c.mf}ψہbFpV+vU\0k{唠(9D/wo}T\k-&p$f,wl_w/$ϟnPנ3dl|n4P~@Dd^ P"ppcY`^1ZfQz?z.>;Dɤ3^+M3͑${kZs+Yt@ K7U<3*k+ĥ3Sq#̜] 6h0؇kC ù֝gt91yQ3⎢:E4ZXd gg/rɯOKsUQ՟lCfirb IlLV+grҿni0Yg U}='ͧ}z7UeWnA\|f ^pI0GnvKRBna}m~}ei_KN܋&EO.ڮ uxyl֋UUGj`ՋmY1ߍ曟lY<[ɫN tZduAY)E?N;Gũ։g1MTPyZ߲T)ci,(ba,ѩL&oN13q^A@Dբtx].Fr.MZWQj7ӝ]Eʓ.#v}qEϟv\9OەO} o_~ZڣFít=x]ߛge#J{4`|h!w7V>iR_uQw{Wz:sּvlKѓM&]v͞,~K~tjS{q;%h֭ӣm5Ku˨6sw^JVũਰKlX$T`҃AgwNװ!g]ZQv1`;JIpW8T˫ë3J0}FCa[-:tAWOu':ԧX_["xvN h0{ {x"׿J$הZp|ngHq”!)f*Nb!3%<S)r˄%RCv''ܧƙ{qGV7u 5dL%Bғ[OfSq:ELx*T@y=o)>ڱ 1>>P&IJ SD#LD)-'Bh}L[Xcu=,a2ԏe>6{˞CņIOġR^r_zn:kAѵC1xe t0.%.SGBV!3#)+a`#@1dc&b<kF Σ*&$Dcs]V1L .CD7dh"BKeV y7I -ܮPd!*| yK|# 3Ӭ[ t2B`ِ1I)dJ҄8O[?dGTe&UgGc?J0G oC9ڐ1 G9c:.zv.[.:TQOba M9jPK?ށf^-uWK`lȘDa 0 ~9Ϲ tE:- r!c*(psV*h&j)8 8'6b0T9ZElUAj WcXɀم75مED$ę.yOeZ<<E,ƈ:I:gI8g4ɬukf;E  k; &CrQc_}ԙLkLŌ;VhhbaN!ϡZ~gC{O9gƁen4W1Qrgȩ,3h6ʲb.Ⱥ{vա'y %s#HT kS+5Ƙ($\[gdr|2z"aJmMPw_*>G W~nD e(gTc{4 /`$a*UP pf`tO.90ߪFe<%>0SRAc9 xO,bZo;X|r48h#X1j?s.m[5dХqTU_YؕqoٞjB/G5a`_>=. ߖn]V`3[yg qGP;Hiw,YJgkSTF9R Ǜ]+𥓶6X[svq]]`D?=|g0.e˘GnVu{arX?NƄ`נ/Э89抻~z]tD~lߩTfVqEX8 ]â~t.whѴ[.*ZQSm /,Z^!Um+?g)ST4PbJqm,EYU^R쾥Ui+XP)P.]4}x!Gtn gr1$ ˖"c^.D$b c-,nz޲j9`NCO]觘滆F9&F$ΰ ĒL3l왃!+v=x2s[pp 1wyzgx&5h''Q/ߔ&cA8"\l8QYN(sMR4ͤH-QnuuNNo3LT1AfOd3SMGf]: .Lf1!;eVKRĥreLtrì-%,cqq,h(8nKck2&QTޣiJccdRIjr, xΓ kۛ  S1 G\\ P ׺m^We(Gގu^VV]P*~6iUc Xe 4b)5t0SRg*,,`˧A? NIeA7:@6M2o+@0L$8AU@pTg葓$̥T UI!l3f1ŐU"fsަ9U^*TĖ ʜy$ugˢsegZ%Y{.s¯_0傈g4'( 1mmrVD8+d4P5(=r4u Ǧ 3LiBu,\Ί˥c?<3BGx&o@G#|H\ <{~I:x[~#Gw3Ҵ3w/oyi*A&^/m$6pJJjH5ːe q6YP9Ni+{>V:كCIDPdf*07un8*<%@&ij :ѠJ;2!=a2PT^bg>b1Qd8wva39˺  >ѣVa2PnQS%6z'+혘ʌU俗g36;B9 r!c@w7/7n2`POa2!jT+/H̩]a$9 &d<޽`GD`\"3nbSTN.@V]XejnQiqjLμߺwU^"[. )h$uiuPrf׏,(br= u)t2:vv[ <7#\dE4!PL%z"c:a2]ˀf@zgAs!fL㇈8~Zo4L2FIlL"DB:r7}8ha2qGZFB]I\mkJv}pyB٦cLdhYIN=.ͤ9>ڇnso t뢓`[CN,c  y+RDH& 39rnA4s*Dmṷs@ Bq~HG~~gyǓMP.@ y'0KBx=tx|60(7KKL u[y<ړ]Ks#9r+؎ L9{\a{}BIQnm;Q$EJdIY$*ڞCϨB>F1 ;4>̶Lմ6{oYy/l{ ̖ve G-¶Duk7ؐ[P1*n% A/3s'j Wף竾&_Nq9ֿ9qR}k9-R|̞Whl9 f=~\׊tp8FGptFib<N#_7 \Y>@= :W˸sk^3i˴WBڿNeW|z[z2W|ߛ52t۟&|7Y$XtbDp G /0$(3:VaRQ-&>#|v7DQzSWd/0ӿ_7yuzw&6+4=#BC/Lz2UecTz5.W 6ɲj]NzՓGUgRF84& fB,]͠- Ao1ztiV%i; t@=iǶWG̸٣ݓa'1nEػդ.& i5vW{r>ˇ I.R̗Iϓb˗/~__|M״0ӵ^~3LN$"/ :$ ;TA ;%EdXJG/H܃MN(I\Iaݍx\<^jλAvɖʫ{>goG)]zvS|6 vXU\ox1ݨv殦0yX`Lq#] p]umhLKo0jֈywլ1ũ})7:t9e/"·<9|[o.UU,ʉ0z *_]xν (JO >pB;3[czJxXl7S:03:k_TлBgb(*LQ`@cN±fE "6#tv1?.WL.QB~tlWb([Z{> c΁T6;ǣzd*d$U頁H] N0P"T$Rژ`>OnⓂ֣n4 3+˸DRQ< 1۸Qxbx6t@ j 1I ~Z PXٟ& I}T)b~u _?j.aݏy6 C 0 ^|ym4D*.AHAw 竢(J/@wRCR6<#Q.:O64 cUwL&\@d08G10"Sbϡs*t$]X_}&Ksnw 6ă.Hkt^dHYx ٯr#TxpZ9[>h)20@08|p^wIo@$vBiR' \&y `X5?O؜ʸ΂` 4ȣDEL8ȶ=Q_D/6xpP؏ӷ+;" NI 0XҵY4 Uvp<8(gFapwX{8 q՛/ OLӅf5c$:O0b ^. @P@zVu$J(FN cHF ͜e?1V^#4 #d[5װ x%`&S]ܢX14;*{Qt~- +5ņdD\jb#܆Fap dz"ަz(>z0{Q@쇝fOŢbQ4 yYKff:&UsP؆S"G@?ߗdH"( 9_x+d(y[+ߓ۷}a]$,T-l?ٴUݴ񫈼D( "Eu+XbXg}m[-W@n݃ c b[=߾DG} ]d[.o 'fAFzI0xΝ=;5[ ~NcOD`ey;ۓ-~* IGli ",c^.DՍbZ{[Pfβ(H憑.TO[g< /H-mfv#7Q /N9< u|N\iCc4Գy `{M7#2C˅!Q*2*Ct$1Aa= Oܖ"%sw~ K[(b(|joFG0Q~w4BF{L_;G5s.2 F bܶp*}bF1j9RNlGJkf;Z mտF3MGS0o-10mjC,8,C A e־0 _&? 0VѬs,~Ԉ'1RYc|?8Χ$L~t#lZLɃ)c\T7jqg6]Rf-e:Ie  :F0,JJpjpv."yܨR4U6;K={l_(˔ ʞpJ9mzONb@G)mGΚht缩 eFYbJľyZ e1&qVlkޚCOFapQ_YLWd6 }5%kDZIm(n6'E8(A]{z5ǪӗDfz'XG OxpMhɛ #!aklՅNZw8֎6SGmFhuG[[6GJTnhGW;:oGnQm'4ÅIz1P mh'ǣ%acR&~ǚ0c$N?ma2T,~SS:w6 % piHی{fނ* fH<$Fuk_FF>v#vaԹ)Kf)w" ]sO}_DGauffa\^9̯nJ1% յ8֗>< -SF#C%|Y#Y0VkC}Hit~z.W`xne>}eHA74HGlÐl"2F ='u޶dh.{>N_Z>onRă`n~ZLT3¨e@69/ PW76c -" ( Nt[`(IGu$id'RQPs~qW` sJ$ȢV6#(TlR hM'hCD #l}ccpu7E.W林6ELaF GS ( Nb'TuSRX>|9ݛ >ڐhƉ_$߭R 8dH6/3^2˟ C 7@/ǍmBKd`j ?N׼{&4ʂxr}Ṣ&/aTWaQ/jZL8m{GS0cqp7}֯|QhŋEd[1ѻQƒ:.ĥI( W O8Ǝ)01hGwJg\PG SK:B5K:Z9o3'QQ8"oFap2h£Kb0VastnQ.p|pHF Y~_6l &'Q6c-.Oxޣ0.ʔlf$=RS?/{6_aεϴ <⑨%O{D=,ɐEʱO\&A` O7cT_hs`#G*M3p9lLyZl%*㕊lF=緃iw)䣄7\ REO_+,2x_0F19+j8$BWQO)Z5)ڐS5(/5Y70běd3ƸK&OByiߊ +Vb }|¾]0F)(Z`llba@7uTM[Hʙf?„z4^DÙ7 TuAC2Ept_riݔڝ^~ƚNǝ6S~-褮`gw;;_A'5zNx%Mtv|j^C ?q,EPw๹fCE3A,DrV?[nf9FP¾-b-O0\۸;)tdQ!~'G>wnsgjjExiLaQE|!ABH4 uEeVw |\ f߱ f+ !@FO0o<ڋ1Qxa BW=PւU+rx؍;3iwx,fd =[p 6&RR 0UbC2}B 04ĐFHbPqWB4|#E9ևcwFUf~)#uY%Y;B2y[{dлɂք];r";46i+J>OJl"Caݟwl1е쫏qعlRfgZd/(zQjU.Cp5$&]o~eLզML6ŀ0R=? !n/GmPߌ^$pK\gnssxZ?v/{s޼w3##~_|}W3˕tB W|Ff#(b tQ {wvC(3'<*#suf&$4(s :=i&+hw$,Jq0C TsAXb&~R:F\GZQeyKq )Od,)^89X4?37M`BQ}kfD栟=nZΠ3F=8k5'mp\T5 :m2xݩpda]gnxV,' %W]} WŸ si=x ݼNEDs8f1YBzmR_03t1\);b\0|ji# #%םf{Y%LdM¸7 cV~DG =M 3:°Ԕny<>q7%'^.\qhsx1#Ԙ0J4](%T%zxg"j#B FuS<4GLaW̽ E7rZF$O7+c&O! QHcRN:I""3IJaU"b%Rmf7b˽oͯ 3}_a=m||)uw+E6䉯S5)Wb*+knUNLySTsɶSBQA?` gki=lnVLB1I1"!FezKql)e S,8ʒ4M?jGEv륭2wNgÞ³;;Tu㆏Tp9۳ysϋ6~|?,fLʬWGBs ];M7S?Gۈ86]vpB ~*.4HWOPym}nCn7K4[v#D 9Bj2#'(zx_CMz^~Xyy55KOۚۚۚۚ?.֤\EvJC\ãPN1sR4k G?sWIB]yLϷh"Yb=Ky>bra;$__ӟӿS5^l1"֗CdU#Dܫ+咇]4Q9R0<.yHJYc*DMY!') c.'Z 4eҀ(nJJp0f `R!V8Y Ey+GG"ѐ>kxm.[KyYcT]9{VPQ9p pͭh E1 mO #`P(O0t\1&߸EEXŒ(i ZPPZI]ߟ[6dX0FI獩h?!2FT* DН0BZGI_tj)Ξ<S9Z0-~]7HSnCf5hpE j,4$q.bV:rx_WB1*FNJBt-qqjFIȘiP 8M,9WC2JF" x  t~<=i@Vj[F7x-ql#1}x!EA+GH^oVi)b[+AدuGqc/72^K ƨ9kucWpO<cT=2 Asmv)?>֪D=2봚J4dhIB f2LMIF g;qBW!p&u{dwlsS)vɖȠc'e8 )_2(y M,(c9X教a'tд fjxbm?\vs{AvYoꙇ@dNq^c^p yh7~W6A1je*0l5}g屢>AConmX=,=-T= ؝j?:viOG evFM~8;:9{8UsUk7[y/Ň{QoySo:QR$0K!S*f15݇2j':y&KA?<><}qXjm18(Ceg"%Zͭ21:q$1fTZIZ ꋓ㗗cx,,u~5EiJ"6Q5uBN,* U"KQ?xq,z}p 9caQb5Rʩ&IdRbBH&*WYb) ~y8zY/ dD3L1eب$LPa*&p[Len66)}~Iӓ7G(r&3QlRjBlA.e2F X<6Z0/q׭[//./ ;v~Œѽߟ->//sut|^o?xS aRw8{uXA|׍e&[DPw,0`X$ ~K(M%$b J4I_ct~8:<|yP{Rj4ĉmX0Ա`2M :N:Jr/_/N5e`zS88?i9yfQnD&T($) E \3à82eu,H$2źճѸ:zp*|3XB<.jiP~sdM d.X)q3򟝼D[h)'oGN)ov-ܾjn[$ugv3޹ڨJƞ; #7LR:V_ѳ,a^mS9hvZ7ͫfM׿u90zP no1 {pw?]/j^+s~uDyƍz ^4۲5{ؒ+0){p^8M &~9s~LvmfhlfχMk![\wky!@2q.| G~~ekonG_%3]yl%Grr&9- 6$ޔ5@d=hrlLƢ(t7~@c/d:^aև$fοӇчk|v5u# \ @P5^ :MȞBɿ;#Ͻӿ@Kו+U%=zi!2]$53&\=k2?,]{rSaz~0Q"a MR۔b˚\4 ?"X"t"UsT[|y # a"gBô<7ߧj-QQyşpJ/QzW[iXQ 4i^[9~W`4q:9,SP"$l{ x\,kș{K}'PYrjߚ1ob4>iW\g.Ep/|p9mNs؇}s,rB qhXآԥ&M|<\O1D6Fu qGvВb=2H 6?3&ϩS w*epJwr2@rT 4.uA=xֆX0SZmRm=7'gVWN?86ZG` gԓ, Y I~̏5m0nI>@10#>6 ˅<$87Yl3`oo C "f5&vaX[Rl%`gnڣa}‡fXmwfMQ%<и`RCڏ Q2+/)/?yr3LEv%ROyiZh0T"׏7RB%3/ Gj}fK? z4 h QpSL,36-C٢3Y6:Åfzr0$;HVS!cO{X/=="m\VӞx]{J#]zA%3Cqr.>'l|A糤T(d ښm7{~sܣs4sZ[K[]{g.?؋ E xo/ߖ kt½$֧; oOߎF=%磕3CȟF]^&]|$ܵǬ$+`XjiXO(Ɖ=e({زQ SqZf(i{_A{tPwOvЖu㭼btھgc5<ۡ'[HS781'[!~LUulUO'WY}"usފB+&{Cqw)r$JDxzh9zZ*YlP^A-K1G@W a,~Z5oKk0TX\h->|B虶~ĵ-Z;TrpVfĞFDvdƔ63{$y-ه!;(Җ636nj)Fkm4 Yѡ6~6%x~Vv;.%xa:*8LXW"Ɍ_ԵP.i" o]ޛFN2>B?0wMW񕢔鋠j5|dza`LXqr}Y*L ,lI6SGӜ]#vF  DҐ8amyVX"RԢ0'@Ѹ0fHCnIgI%H2Ǯ(;l0e@!*5WAfܢvζKM\l{En{ٗ]ptdL\wg!|'= n莃FQO'a…"}n4N0;AkeE=p8)vv녆n6쀾ake.~ȨcY_g :q=nA*mչeDn ʩDT~nsv@1P׬zoSWJynz}zp;ԝ·x2p!QPTKST=0ULSLTL0|T|-_yĴǼ/x"@6vYY>͐yCy # yAI6(0Iˤx*vm{@NeI-I}Ӥ  &pY Aч=`YtBe!Ł`)CC? 4~ORP(z)6≤K#@ӈFFL:kHj0`6wq^.>LZ 1=YL;`DVh0湡il42v M\0N_DѻfaHdN`xeV hwMڶ]f 6`u7~nN/^U2&nXƒ a XeX猄!]ΰm4䇾EFROszYbR.xgae[b6v)F  8pHyhPl$ߛo޹<1 Xp k!+dV(Asl6'[ȵ}DT?Hnm;I"C> c"8&rLNmd  1BEĮq9mW쐀0 m`X(@0&08 }!]ؠ36y[m#[{ׇ '4H4ߟo=DQ]ڽfiֽ nk!G@VIvH`*4tzO;'l;Wb3b 4"QCs 4c$l M n {d UyH| U@GPO! ,& .Ԛ@El@Fm [#JP aV[.yY9F@5.c`qva>8,?tGX)|JE^ؗ6 0.ZŭY5ǭ#>a_.CW_WdM)Ec?DDWE WgB,I?q,:*9JK/-؟ҏnվ{6 jOx7iy16rLKɷcux1{"I/A[lpݗ6tӏ] ~7G773/֛/wDgq"W#!c7f駳ox1F+fvlb^nڻ C.#ɱ\Wsd e]@c</$u}!Lr;1 'D t|bڮm8 b|<r9{y.NEq\6D*:DkLd߈ڇa2vF#ö.}6O("O&GMƄ5n:I8 Ecɳ8R+axqEtR]nߒ͓)3r~9Yk"8ia"ҐJvW"͸X/E̍]^`amץ<]2FN3ҭt!Դ;U!]VK_͒tRE:Mj&wl1N`@sqh-QglFHp֘ ? 5;9U+gd疏heDC;fǽn Q}oy]MM# 10)k6@Z((=/ۏp؛lYHKXyөrIYHIY6`I8xνu.}pmp6:-^6d=CdsU[aqe Y2{,%>bhJ pY 8J%>{l$qoUE:[d((-|FM[VgMzrlN)ȧNS@ +S@:#!YTHT5U AUCPT5U YuӚO@9Xp +T&ԤX"/GbDHJD!:ʐyV2] ]N ϰXZ2%O俇5軖rXkÓ* thr]`Np;)tJ1r߳I j](@-bS~9l l:1gZ{8_ƈ[w%`,65xM.iqM̐T Tσ#5'6`3ӷ>Op>JD] jYߞ]^"B GJeFz5ߎ˜34s٠lSA9鏦f qCzL}CiugWvmXwGgzV ]gvwwvR!lUDz[r+eCclcK0։ߞxsvuVCщ6>(`6:kQ)ƽX]"E*nV>]nLGlL3ZQA]tFEp`>ޏD>ACx\~!S7k?ph1'|–b}i߾MT2ooUK$]"HU&dv2jBOY>'N3aڕRZ:ݟ{(w&:բBl%ƋDjU|$"FZRb)XvQ%yR)&#CբP!1k-bm*cJEU`yQ\̏=)N`KtU)/p ͉kE B j-X%Jb4J(ԭR$&rhڠ/߼܉\j\7+ldiuUXhDIFRȥ}(f,cSfpf#FZ+ HM5P)F|{2Z" ӣ^7gK#5%%LEz\C۰kǤFid*)5,:@r{ xB :ւo;\Y`AH䐅h D]Xdٴ<%CReÖ[vңYSB'tIJѺƟ>(:GIWU^ԌS%XVωQ,Xm4ќCk$STB :j*Ia~IچFrDD61T9XWV #eU`K] #$E~i!MՕ,y\*$(f,I:gY笃 3bhhU#)li_(KʍQc&C $XY)dS! Q c͗RlۄST"z˄wh|"=6XȠS "@XVXU0:Y-w\ u9kx9]Gʐ~4`&TCZ\ E#I cHmV\NS`Q#etƅTCFKjk]Q Lu}Yאk$Iz ysaRA0PVC֗]b-`P&ZpU:-h`ℾf"]2UtX4d 3:eȄڟc4.4'9[.J2z] v>,T +Ԝ 5h2L[! `E͢Y aB@+dSϙBbYLq>B% ir`NɤI"]+ThoSBLE8#ed!_ RY#kG&d)$h~oA!OM JQ]Qjuʺ@J2"V_\\쿉%L0K5$^U dZb4hXy@/!$=—Te 2jnu͠ ۀ8 UgeSS_ }\EYQH 5Yr6e5$@d4XE hr{joX;1>9ևeh1K*+xm c,"{إ1=Ŭ:"JL ZS `"zjI7۠k@)@2"=N @n3h v9 ѹcAZ` " P6vNY9+ZMN7$- ̓rHD xdI).`1mRg&2S"Q\4y㠈@բ`&yh#bl6قJMʾC:kd9Mh%r[6 iVjJ1M~뭈*6 Ԓ ߬rI ucZ)a ~cȖZlmmK`=A;wyy lVmˋ+hht0ufU::nѣйG9OcGŘŪmEXcV^k $UGqI6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauQ=&;^7M:@`gdquبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6QuFGd\wvEV;D^dZw '$Xo|{s|Mvqy~nvҟOY/Ȕ48ݠR Ee "jWg#N+*1\ :~Ʈ&^~;^o?Y+#z6fc$ ¯/o?,~7@}LA䓗!?Y~[.ky%|E uMȑ h$Kcmc͔o=s#?v̋Y^훳^6}؟/&z"m,^ś~ 'S.PXo*xv5w*ͫNTG:WK'TPS-/i,>*{qV=cz :'V!li &i{*8HW+.⏐'\2>4{|ΐo oK6ZuԿ:]@.-Jwc4t]t.Q6˶S@W~B194Of!>Y$4ii|~#-Ii:;)dlMoMvZg5IL['sF]\]o浼gzjbZn@W-K<=BY,qD?[ĩٕSz;qqk\my7aE{nCƭj(:ʾnRYNԳn˳!ޮ aCfEr'Lf+K6%FQf[{_%]q歺pn\cLr @n Ew[?TC|G94O+(-ẉ_v1o%p!G.ϋu>Zdɱn$J^1$z?;ʚBrr`u?b t}r/*by#S(ߌ;9d5yKH }QA3f}}2Q!ѿ8{6tU,vtEs eh.IeB#1`.Lo.>e 4c!CK!(.[4(F1#r0@~16i+Y#(kܒ|t6Da ȃ^&5n]ppeZa=ٙ܄-Jz:KhcX`tn|Ըxa %+eaB(8QhY\j`ZɏP`JQZÂhq,xay*M1B Hs#<{|~0ͽemPH,hYqlt96;`2@~AyQe-(W&ct" & Eש?O~Q[zRyLK0X 6 4vX(A҂G3e F3ŵegxmr`cy-,N>I%ܙÄ4&Q12שM!y3L42PmeQcZ@IH"?lyc&[YBo#ǜ cUF!ay+dVb1b@(T2'RlP' lьm1 ;t3&ͥO 18D>@19) % W!05A%Բl}+$O <;B ϷzbZʚѪM&w aynK˜\t@&Ũ: N@3B VgǴC3bc0-ѦͦM1JwJ&m6p*zZ0m-JsS>c, r6 W1С,%fcNr c2@~䅃>:cN.DɖEDDN: A-1J_I sYE 2qe\va1Jۂ^]IJ+E5]u;&vu%0_]yWWjS6+YHh\J'f:A\l;0\!`X%"WvrE(]\9 +u-rhS+Bi,W'(W[k"BNד \k+B'Qz1GWoG`9GW;إ#׮vB GvCٻwr;r VV#W\-rEh\Jmf:AAtU$W@jjES+B;ȕ;Q\IṪ'"d:7uBrur@X*+섪FF|P9nK;pvCn(Ģ+\Yz!V"BVjW\!Z7}B~j/r*rk++"Ǿa'/<\%YNPik:!?ÇqED#~ru%uۀ]Cn>%f^5 bpP,7)xpח8\^e7<.WbeXU^/3b 4⧭f^yg+m2AnǿZ2W: d\-z6*`Y=:zւDڻŊrFX/2s:* \{c7c`oh /[vm_u1wϕ⽹kqDD0;J3Mn\4PwYwEp5?/IVy@.oi]x^->+n{[|opk]s'q(_ |F*#,ý˟h}y_R]<`Xܯqjnʫ].~̷oE3״v7Iݾϋb/;_m?col~pC-r]<-gO]pOлļ|R1HNYhtŠVvogB;νls{[)ԹGn>>R^ȸQnhr{1W ~j\A)\O\ܲ%JE07M!3}}!{Mԭ)F30%yüՊ%GAo7Ov0ˣ7W\MQ 7,sҿVC)*%#bxn^:拞A>}1K~h\mryR8OW {wvq/U u^}DQݕSGC\x 7J{7aEZۄ۰ ?zw=b WcgMA bς6ffۧ2=D4 q`i Lo|AĈ>*E\wUQ!Uߨ0$!z7ӝ6dL>P cY_Ƃ}0ӿaT"1߹gD^%{sC0DX!7x{;6nD  po'Xx,i*n*:+%RBү ~VJ({fgsVP;zYDW.WZqYNG6^h++z眸\! ^7tyHDWWV Zo.Wru:rrW"MZ.Wh.JV$WXnp-ER0PvR#^s+?~t `ص#׮vC{4;T;ȕjߩ&P\WWJuP\!ZPNv5իXkHz+[OtEh-L]YNP{U\!\+l-rEh\ʩ=w5իȕrHW\w~ "O]J;]u֮ nz\Z|tE(\:E2)D)\YMЪɟ Jf:A^ ++,}5rEpEK_^}\J;'(W[|Er-#WWU#WVO^\jCrGN>OԾ`,WuG~a7HN(mu$J Wz}^p]trEFzrE( 1W$W؛j Zk+B |rա\\ItErEm=rEpEfrE(f:AfBLy4s'ZL=[OKsz]\ы^4b=wdi΋Uv^Ap*hnm5 u Shŧh4zN" .Wwo,W#WF 5=~+ES+)s= ʕҪjWq[\\frE(\:E޵0m$鿂&wq 6k\Wqj-_/\S@TuA%$%bDEo{y9DWX ]\I]ZZڈ̺B]\ 58"!ҕ>Uf;CWW >tbp{Bttute4NCWBt(EJq,ҫt; m+DY;s+Jz'|w{6fh]5Dٲܕn@WMUOP;DWR ]Q% u.PRN::@bJk%+ !BJBW2vBm[wգ?L#e''YZt--AT_}J2oӓ K6=ATz>{t;=|%5"}Ga`|0'qĐP}6v|F6o^ ?Q1s>KwiPJJ[eϪ{f WPZK K|KmT õ?zۨq:5+}H*?ۣ{]Ip:Xm^B]:opW4ҕtuk$/_D>k*7=aPDQlZzr ^&Q]ևI:iݖ&P_ 0YZn pra23z:uUaIUϞ7 CU"|?0z\$LV"Z}Mq}ϝDT3ųjk}&)ߞ;:;y . ϼY$KoWu˯t-۽&~u#=icsVi-DqO/ywXv﮻4[lw9X: `ܓP{2z Eʞו=־K/z:_Zo_J+ɗ\Ȫ*? 2zq>9z]3ěBQU*-6h҆=V/kW2Jj*$Ե&1܊V_ݺ~ ՙ[]p[W6ꊶs\LUn^ڀc~= Rd6Zs9N;;2\pN^{xalWkM@?Y͗5}ir!9hw gU0^,, Z‹q* ߓ"U}x;7%tUW6f4FLNuj]x%OŤ{gi sszn5K3X!q{] hMZ_=lji/OT%uY>Dh-(?NYtꟋ>A8`R KiMqŁ W(<i҄6q}z4Rxl%T|/Syr g'ݼ,a gɧ e<+E;:7e8=NWqRskT Biዼq2KOaw$Ʊ,/!W,MUL7oӏivՏa^mYɭW$?+fʳxA=^FsBbYDGJ 2^*UZhjeuDS>iBg|X11έa .C ͪ4R[E *N|(ZdwQҶkaE'\þugz֘"B'tjHo]]ŊkNJ vDMyra5G{%l@hu $!z+2| JFUe|9 |E [beBΤ7]pvBň(NF$1BaXkΤx6 X0"ɄT$8B-C5 ^ᦽ| spRk擲 nmiu7 ڙv|.Z|48q. HFۑj$F@2b$!ƘhPL+MKv.麡uJM**;Q]w(Ǐc;0Gv`٢1悺t#Vp3wb!Z~#'֝x@'9ja.[3[JךOb̖4>&s;`AC-ax5;\u[uR[  rgnE}V?F8u::CWWW ( JbKta,]\f+thWXд+c|٤m +k^Ahge򭎮QDDWBBWVضT'CWfM՛>1Jwjؚ=U#U3lOU3evӀLGWZ5sBZBW~iK Q2B]`ɔ3tpum+@jwwtu8tŭB;DWظ3Dd+I1!Ig ]!ZENWRwtutڥ fpLڇ^=]!Jm;:@De+lv.uf0hl;]!JpQ;%{_ Юդt(m7]vt)}+lݡ+k)q-m+D;:Db_}#vDWvO䴣+.5.EWX g  ]!ZNW.:Dٺ4v-P]IvXwDZUDkkIm2fjC"!9!4) %Bũe `_]Gi-ft(WJ{oP˖nwC4~p:)<i4>&iY@I9^A16ND~[t@/]JWT%ɣ_/N\6211#beL2Җ=;}q]n͋PL0TP% ˞7t]s\<pax@h)]m+ AS~ fGWCWJrR+g r&]+@ +F+::@X,1 rg JvBttute31ZP/BO!k4rD™eV~T9Ժ s]\!+thC'}+@x?k{'T4LɞP6UвeۖtE;TTθCt-w2.]+DwmOWRR [!++t #D%S] ]q#/lR ]!\\+DԶ+AӨaZx]yu1\if2Gkfvxg\9LɖNɎ܄DkϵIJLc6M*J3^,W=Ë$gOPylp)H89dѩ.ヿI5ZXdDbaB'ZHΔ X0"!!"#FAĞN 0(ƞa歝w<Iz>% |߼rkf`2/NOY^nTG)Pm &Ϳ_ǻL^FW.0OP R`ʻѣS!Xz`<-I9> Qv2S??TqJ8A)H ȹ P#JF1X&%<ʊ 2@cU`9V[8uأ~g22*hy_|%_|{__}/G;gx> s\Dɧ2|""U0 *\B(iM ZwzMˉ?N1GzީO/ݾ>#7o,ӟÙBHITn^Y<)ڠHz X7zQ &{2wy^z쫝lxWIة^7#ؼV~LzL`A>@Fg geZd渦ه)Q?῾}˫o܀'?&P ߽I>|#| >gZt8 avN,?APO3!)oAIJԒ"LlOWݺ>M1̏Wum aǰwTvDto}t_>n쮾x{__yQe/(ugo i9?|G(t{}<,,Kv\xQzƳ[m7\Oz˒l8YzQq4/OOZsɶ<Ʉw%MG'[kE{?# v\,Bsr{  µdw{!>Ÿ] qх⌼m:$>ޓxdz,dZ҇׿{U`Jݑ2D{holo>͖;_uc{sUF>9"׿#gG]sb{Ӄ \(-k9\[˫n~ >qBz0L?"+q }H}拭 Y޿σw|ӛnПp/'OmluF >JN6esӛvtCЪ֓V s2[5DQ8~*(ګs{{6ǩ=8U]V-O m9ٺlj~]nj),bI@\n9d~{յ/prsHr6GlZ#1Jߦu/@7MPLi~c'IsvFsl`.L..]7jm4c1Yl?IxP\T~6oJT7kӥ͇7 ;GHEqE û8H|}sy߳R^3kbkvWp/A(W] R j4"r2Eix2I*ΡS_K~ xj6N5oOru W#j1ϫ^WneYAKFE5LKŵLF[h!Bɲ yHRjBc)KQ%feلt`lIj.? ZqSfLP "T8㾔Q1' J.pe< ^Gc0hѵNz/ 4?:hT)G!qάʢ\&U^Baz9В b:hLf cpD38fxIԚ#*Ċ5ٚ \|9ӒNg>n6t67JD<Ɗjј79k X\LYeȵ:[6(kpF'̕9ƽQDl?5ΐFg^2ip/)3g]H\$-q5n}DGdI6.F)['OMʆby۬Jܲ2"JZs9VPJRͪUc1Ɠ \Zṗs]ѻ9GTSpse_2`@jbaz=&XfbBVCID-Xʂ_*DO*RbrilAgY |Qb)IX>TUf J} 37mQLȃ` :`ԞW"P= ^m..9aJz)MV^ul)lDma xLs v,0:^-5 \ ԁsHSYBy&aDUPd X&ܺsH28.@s@hcnU&Iq,&bGYGqkfCJr_ :[%jFSeY\`+*5? X-Ҹ1kP s10`"!KN\Qs( J[2JPXI;;&؊~BVSY2Y2V(+)2$Ȅ1-cEB>rCAJ}a<*#2 sYfP꠱tYD@1˥tr\F/`B]^j>`" T ˕3VH20[! `%@ "t,:\"B(,( Xu0wAd $UMɷ%;ʩllJAd`5ve ǭ`I p2M BPUv 25c`(qH(ʨq`j VڳY$o(Eydͯ(_u q"RwDbkL]ZJD5{TR1lB{`B`3/=kG(PH*+9JOQ#p l-@x@ItvPXoT+1sՕb1A t3Vk 5kCBňdbVGq|M1|,5j9r: ^L عMgb{7VZp 49ac[-kt2AT瑵| b^x1Hu`Kysގ[БJ'.  -*=h`j1xLqΠ&aB|QBBGą1 @P$bDM W$2+BiUc4nh=dhEE@YR:[Zǣ&P܊`m@mBQ,TG@''GbiV5SRe%) CdD0XAOSښ~a9TX] M("f_[b =ND@6DρK z1΁6Db6y-e}trQV]k@)@d2+"9.JfR,JLT@$4.t-c|gAp^%.5ʾ́;J*!56R zx,~ 33A#-+!-t8E zȕAB~)Q#2PG՞&3tQP 6 8c$@ }KN JaV;ͨ1rI"WŪRDL0IJK1:Fff%d{lB#("pĒct\NPklץ5]];gWL0P5aWz齸q6[vy#co+[iS-vbr[Jp3 Ul͍ۊuۦd%$(w:ݽoڽAO)& ;{{;/u| Hm7}LKimhws~/g}_~I5m~t6W_z}o|Eby_P Eg\7O:-{AI't{~2sz݊{7\ƽ/Pv jEqtٟ8͑ȻFӍx4/۝rSyףaZׅqu|ہ{?t;YJDw3~cWx ^gX<"g/~)QsTW7/I?wkǵ/ow/m;'~h}нk^o?jNw9{S??n6Ýw/eY0TEV82h$Lm۲1ܲM8V6&wAO=<;8=;)@\Qw|qtlԛH_`ГҡSp~|pn\^ֹړR4EX6#< gtEK3m( 66Vc͓rެi'6Ahk6InK(4 J;2)3)(` Om49n46PBq /(z#v;P 2P$"V" ޳x漹S$a9Lɳ?y7?ai~`z'd|"6~9fAV<n5 Rz{R.O'{;F%eMr-^LFR825Z߸T+Q`ZsI㬹[/"ͯt~qsh]Al|Ӎ"L$F7H G5k@BAâvKY<#^,xyj7t38H.kWq$BfW̥R$9p%*\-jz+A8\ʎ++bUsWT2Vj qŸV%\`͜5\U̥R[j qeF9+qW(WRWpj5);PWM/(3㡻;$1].t!Bwͅe`OD/f;Gy-׏GQ׽Χvl'BF'cPԛE5W̩V)Cp6`/~G`ADȨojKJhdž{x1iVE XVXý ;]h7᫏hl.Viޝ:Oq|O3:t5d<5`*Vs?\b^*\SK+WբD%R P.3 ^v\ʲ]Uz\1 3"ʥP-e+cKB:+k]Z*TRZj q%%U+]NX9πqUr'V˅r&e+2fVu(e01c3M[vyԿR[K ԃ̪?:] i& _1|%5/@1e[Q@bCBz4Ѩ3Lɥe?aJ0}EԠPK K|KmTT^m1:~q1zP>ak?i >?i_3< ^Ihf`UM@70ywl7egxfE8l0m6Rm܎m7sBŒJ1Q|zA|NIv7@ >l{+B@At׻- ]?keZM 7=F?ny_5[q0Mdmƽ:-ƤL |7ռor+rjgIFvOIvnnґvxvJCGKU''G& gm+I2[#j^?=g1yN`ds)3bA+s $hn!zkY- Ͳw8b&W u=Ϛ;MfEsWTwrJUZS0[œʥ\Zˎ+T Uj}p%j'HgprѮ rUv\JYjq%TR:+ AU*׺+P)+;P%#WiM\ \\!\]J++\!'spug :T>–WR Wk+͹!\+˸+BBWR Wk+å. `K3B̙AT+JU-fWkzW X0wpr3BSR W [DW9ଓ*Uʧ`>dS:tEMO;+*8!\\],`ի aK+r `.Ȳ UҪwB+* \\]ժBFUZC\ btFy_,,}p^.yd\*͗L7gz~ϐo RFm \o,wef#)^񨂘&ɧ޴\ F?O!^hfaN52'(Wl{d\tgsPySFw? q۠uIHp}6ZtH u1:&9բС 9Q ,ӜI-$gJ: QQ)B9+γ:?%pXe>ܶ)1KFW(EoЛ@x9-/k7,N)$PfDŽOPT 9x H4rIf/WtV}U7 >Տ/ ܂ l1svmR5Rw[j1䠭T8_۪gbLn]}z /Pb<&ap}*X[J}겺(Rf^IuIy2ʢ_⅟ݴ[v)-$H I-8H/K:~rzN|YyM-KӠ_ TTxһ,Əz w!_:(.0+ԇl}[ rKo9@˗c+orN CqP b' =@|~w>sH6N  |0y5BDt49?; s?&dGr?@w =GzCչpYo4ʝM(SFV G4sKI$W$^9kRRƹt[rRHB[[ Q(!gӌZc/Xiϴ҄WZ+~M'~ V]Gw Xt0:YdϧҡvCk:%ts{@5漄n<՞ThCRtDIt6Z0 u__wku.J+X%1񭗦BPOUεA>؟VK\'(]gܭ-L!:ψ 8**f̋r )Np+J2e잚\&.&Fr.4^d8*LZMrmi[pV͢5RvAj(YrGaWVW}ƅSZ䫵k~;{7&Õ[G*?pZg󭣜O*r&pHT.Ny]dR { f@X.hDQB9$Å5S͹ۂ܃ 9#/eJ/$v+YNiGnPDqjZAچt؆iYpiInxM3o!fy;w=ϱr>IvP xs8x#"4~*͆"0Ib6D9ac?1IN9Z.|-~|_xy5.\,5`4ӷV{>E[-e0P6 |u[AȦژSܔ ۘ gtc2bÒ'7ܦٻMY*#:T71ux6 K0 WdX̯XHoUdAu;Qo^_՛xJ` H38DikKmGG}tdilxlWaO-Pe:<5*z=OszIӢpJi2Q G'qϡ*m>XJo,|0X%ILBMjnv9M, -EԹ0<\I|{%QǷsǶTT2G fWA0%veP߸(j%Ӹ$jlC]%OLjwE{,z'WSݳPIV`O3}4sC'"ˍ%6h|fmy/ʭ̨2\M)׭B>J-S`WSmP J{/tKەg؈hD"MMQvWޓRr&̓9h١momZ6VAZZ<>wR+eMgiEMP%F"o=o]cvuӼ"\}vPlӥMgŶ3n÷z%UwYP[>!ȹv2@ AdL ,hNYfdV{o%նv'<=oA3ϼA8I:O}eq]b VH,V☓V}UnˆT?:<\ZjŭzkPG?l bg BHJ&$hDp2 D0/Ѕv .PD/L8vvP]r UhuM;o^7ieє=cI? ϪwV 1ޔ*xRQ-#bVX<'/#\ˈV3+T˧ȬZų3OWQ ]!\.c+Dnt(tu:te&BW5%g-] ]Y I]!`B&uh-iIі2[V%=;Bv+<+Z}$ iؕفLKWz S1~Ptp-WM+D)Z:E2DDW>"C=nhZOWRN ?f5++e,th5i:]!JC[:AZ0]`NQWX 6D¶tutH]CWF&t(lJ.uLCX1 ]!\b+Dԫ+DEKW'HWHidDtg!,d@cWRcWHWhbreBWd+k1 [Ke8C툲]wѕݲmXN8:]X{!npvBh@WV=X'c<"BZDCW tBtutŬQV6BBW ݟ%+A`6"B6gc/ M+DiZuuto j{ړsWwG^ T3u#qS2C}^ɘ^=9k ><& <6;"|FmvƼ{E JBTۭf?dO_@JBxNǣ~ zݫ6^kI} Nj)Ro s2X&x=.ʖoby65=L%;J4ߣAVF\Ah-Mh%Gs"[]` EWDCWVdx)ҕLDt#ZaG3RN ZǴ e8V4-] ]"]\ m+DtKW'HWVHEcDZGCWרX j [V=-? >npNh !sGM+]і[TeDt$BB6{-]]1,gV5t(mNL sƣ++l,thm:]!Jk[:AD[fc9wj5_.CϹ;Y[fJ[Gxnln+#ER+e#hf$rGhq79EFJMmL[q ]!##Z%NWR떮NVt3+ޒFjvR Ta^$z>R҅~M}w0y}vfy/`s1d6*e(+~T7'P) ]|~ep ?rè ~:`,u@UJHW,VL!^LˍMo6ζHpf1v}FI𺊹Ӥ(5:{f <ڤ -eRwR&q`*Zs0 %73PQ{?X6f|D٬#S>Psw ]wtzó1 plG`LIu0s&M2&йw!)J!PQqpzB,-TnE0E$NrkÕT&#ౌKkb Mi2 j'4"sN\Noc8_,g>0$W?6s;j66C&%% qA(y.tc" ؋5n/X"{z(Q͑diq-auuu9=}es"X"T\"Uq&2cZkK Bc.:Kɘ%sdxTV.ke:u{>7rH5KjrYU3ΙUYT]`22IbׅPcM"$=Xτl43 a8d<|#cDJ*|M1flh`V^zYuZyh`-n6tҶGL*!M"#:X1U0 +!Ck` Vgke ltb~\-"uڤl(Ax[U[VbUFBVk< JI @ $5-+T^@T. U U:?lq@) /. D,Edu}rk/́uP'SaA[w5X}m6F82=bTnBl \*&@QR"d2;ExG:&oCUQXÎa: L1\;djy=*fb-ȊFyN ВH۔ YcvXT5I8KmF|d)e$bzc"Õ,Ƃy7* s.}^8,*VJXO'R% fcI; 4]0Q= wi`J fB޼jP gUn2+{уEXY oT[ 3$]$y4$oUBeڃ!K֭hL6b1yڭN0:]y,<L.esMc&AY"k W@7]pd3 L[ LSpR(uZ5$ў繨)UFs7zS^< 푾0 |ျaQ^o* 9QpyX%\D[9DKh2R ; #*|DHz;`wƭz 떍M fV"bʅs <@bv%M^gڃ7FPٗ9^u$k ˚ÏrfF8ae%3=f@D2\O#0%7aDF,Sk2zJ&`َ\⌙TVo/ .Z6l!&\Ko&b!] 124+&5"0"pŒc8.Pklץ5=];gw,0@5az?,n\,֛0od1N⯚:8t71m%8w{w_~}fʑ:nΔѳ?V/0-&?7bsnHr rr}H;٬~Wbvɍ2_O׿ln~{л/'5t^_^raO?۱{VNy bm.K\уc$A06P9V .IWM 6~Jz&nybd#EBQz(Ŭ=k`Q;y:F x78R?O~"HD'R?O~"HD'R?O~"HD'R?O~"HD'R?O~"HD'R?O~"VaZ|307íоq'?tVVe_w\JZ.E?VDCyޖeo>hOS#\m&AIZd^MqXq# ɎHggZ k`-g_5+wqߎ [fzEM_Kn4[7JQү\[t9[ou# #{}$7A8 0߯;'y u1_ή?j>c$*Ҡd5y7]M.!~Ըs?wHC1;?7S=IQV{Wԍ[ TYyg t\+YY&}[*d2JZr>G(;evx7ߝV!vlc9d0v?u/zxjQIG~tjt}= Co\ymbvy^>v7 l{ӫoQw{ӆ?fWӫݷUM~IfdU~H/vBz'ù˛W&gL^G*z#ZsRѓT'=IEO*zRѓT'=IEO*zRѓT'=IEO*zRѓT'=IEO*zRѓT'=IEO*zRѓT'=Z1I>4̕##HsLiٶM2}y vSuA*-m+ªv>8)t ]l:vRn,$~T}y56d^y%}~yh'yO<fIYڀKsaUv֓Obw˟5Ml^^̯&VSR[au_T{!飴t ģÅgT77o:ȃoT4lyv[RO?8A02;O}N{;BaXAlYIMlg>٤&}6IMlg>٤&}6IMlg>٤&}6IMlg>٤&}6IMlg>٤&}6IMlg>>[ 17s`V[c:+m7eݽX6cQ,pq@j׎$=%_{ai[ǯo\$U/5o֝Bؒp)&15Q;{F5mZ~jPoQq3RWԭT$P`q*AvQ΄OLh`D/#96fM}2na7%,=ۖM?|O/>eV=r_%\/]s-]ַ:=v/{SNb-hnB];^i˘W:^Q+xEuW:^Q+xEuW:^Q+xEuW:^Q+xEuW:^Q+xEuW:^Q+xEW|FڊדI1)S)y9?uOkuL})=0RZ˾3K-0RlўN/zΪ lj6g#tVs)] Qt mj1 m'hn,1hgb@ruߋ)ׂwmsTM :]L=PeEqsΕUYmS3u݄乊(pujxfԥ H&K@$$b=s'AC#(S&&>dDLuZhj2MC4`bkʞ0d%RE.UӢTDzĵe2UTT%ŢDL%jѵ4C2]׳dRDDZYȚ65f42ښRoDN ~ ?;.=eje,8=?p}{)KId2 V˷pc!hе){= oo j(Q=&wIucs` 4QDIVWlbxЖ$ [I E+~:`>ˑ`\E h &אe$׽T'(E,L|IҏKF\b}#ބsvM6_Hr.g׋Oy}|%jҋZH &\[KE'*MwMiеZ̾@W [ʶIc2Ч4ɊK;|0{g*8A!} ӳmeD-TkK,\WE%Km 7GtQ׌\ی"KTea( r5y~"VNֵZWYi+H%rLE1Ū{ _5%(;J5S21ׄ3WGͭgΉ$'ltoyHHYéP)j_{hژU^Ѵb4ʳuIgv%>-u+@rLzW'ɹތ.˳i_kfa] "B*8U2F/\!}ғtu"rk# E8%DGTGWaqp^` ?;{ ٴ5f2UqE抦h԰7>vrE?}:C1TL_iiR^FAOi#BwonRkhu, 5M`,HyEO#H}1)bmb.1 6xJ l сkǪ%xtYg =E^& W璗 Uէ 4e^0/aj*4$# ]F=$#J䎝uW8c$=IӊfI)X&)3Ha<H.AS"BKH@R }/];IurC!nq0*3t供Q~C)RR]9Ixڄ|Gĺ_G mpYG$)+? N/NxRh>'}NPTO^Hp437tT[e!%ԒgRvэb;δ4_pp&bi;=4{ޜ#ЯI8߮rnW~Cr‡7IgQ ~?i4j'oz"hHQv̈AKzw& #-?qpq?:s@8 [%}RisT$ 7on]G~J^|C8}}Xszs )o)-дׅE#%Hٮc6F&ܜgն>t8ꝞqF4k^WJl0}ƅs֍_3$ e޶^l~>]NK7kȗ/P?V3Z+BOfk|Vᯢ:@7z `Fk 1XUc!0=/ig{`JPo!Vr3ԩEQ/m;XvǐKY~z݌PTW@f}`,2 P BSI|b˱giavQpf eT۸EP8>F7#cnZa]{+]v# |A#y=8w +%VSt2 7Gdޗ.9oVwLʸoT̺́\w3%_ԙi-HxYůaBFNwZܾGSi-/l4N@`ыQ21NN\ѻ̾i`^ڔTW{>h"G՛ZwݻSLl i|Ƭ.DU]8Weۻ,.蹳^hN4ô3E7o4[QW7Ӥxs<^7% E2Tӌf%C]X< W&< }Z-ۖZ=YJ5~cB2 <0YRE/H{ZKp_iHTkV&dSk B7^hP?kXsxtgܛ^];ӸK7DU|6q3@T;O5'2j_7,ekb?QAhQw@O-؜ȨI]*+e "قkٲ{Kw{\MdCz a<11]dS4SsURڎcI$]2r@л_g^=$^?}Yt(:![㆒Ra"dSK#y&Fr>NgKyhfYȅdG!\l v``={l&sz[j %7G[(25*ҵ(˹Ϙ;!iZc ל $(VGMf5f Qdv6p^v6Č%Y?tFv~Xq"=;ybH!f(/38 1x7qD%cslw,}2@3Bi{Q>t6E! `G#Lޡbf2 N%@ 4+:m1kНCRJBآs"zV.Ňw~ Cd%wfb:X}Q HDbڟs5>$ov]Cfna %*1>)>Uϊjּ_\H"Z;0{zە"P/| w *JZTRbҠ $C $,}rL\L3y^tɒfcp8^JT烬Z5VmƜ$$$8/q?;iwq^tۏrw4*@'~/ ׻u_?~1QW_]fm..Pl/ -" 椩b5HIڛu&\Aw|Xjzbn _?z.?\wy[3=Y@ˋ˦u~6=-{V^:07 Q!օWU+Hj3F:UM̖T6PNhr9nmm_Wm}%AF=iR;ɷ՗6Zz1~AHtP2D<*Z<4N@tVWG9MxiN e,c"C J!B1ǕB9xTƺ҂PlBM[9o=va1X[A5E%S[\@_I,{RاnE͎8Ύ Nxj^yjc T lCℜ$S>&L^&FQ+eR1EYJ;g"L*H_Hoo&o[\G_rN-*BRrQ?APCV<՗EǸnGq(1@`dp0sHg[E‰uX|: m+ftPbVQR j|sWL#,8ZI} B PeЬ.x{Q'jVNۙ?( J)i0zDif #SsWO$F6h " ގ+<_(=UeV"d?6 "< ȔAY? C\!xt/V$L?wK||8YH25ڳ$me#/oXOoãg8egVESUVAc!$Ds5VWSRx֛_, ,)rJ5AzX5r.&;L66+FǜVw~fѧ~o2jr(G9:&LHes5'Z bȃɣ<yV+ٳ>:>nZ sW*"Mq$6#pt-8ou@$z[9ݎ0V'mS\)$TƂd"-MoAs"|?7BqurhE*L<)y4SDcB$%\2NDdS!l΂wDðq3τ3s{_8^dcjf=bZݗ(mÃcߙݿZZ|l:]e1cY~V?'c듎c- Y$K6~,>{[n;LL(3F/aN(E3fͬ$C~Ze&jԓ9DsL(QmGs'Ȱ, Ƹ#aREbLe=m*Z AJn6qcxZm9&ѷGN@6O'VZLW8G=?\Jŀ`3m"0?!m0bgD@8I:-e#f1=P|ovW.Ѱ([2 %pُH v Dȱ;RRc#67ٳD~Š Qpv2dn )j=[u\k`_,䳱5aFˆ^FcD2Y+냉ԃgDDL ` Xy$RD~qڟ\܍{oyeož*WT(MydXZyqD}T5VlmT3IsLl mdx6 ^h3r%b[h)֪BXZyڕveg߂AoFtffn9aS ,^N"'bd"oeyՇ }z$_Uw:/o<9݊ x$ ) U2LwZ D佖豉S؊N,'s;9v6yo؛^:CR5jxT~8!xҒPNbXXH&r;k6~4+f+A|lUMi)!kߟF%`d>Z7`G Vp).}Ԗx93[{-Ck]5kѣ4"Y`ŕL;-jMU 9E#&h{Om@o>M ڢϖFU[y%@8 -iPv8_}'[& wS ZG}~יܧ3G3]0۪]ͰuO fb{ԴiH䴎}f8e[}Xk_zd>XEnB<i- \Q)q=Vn^ڢ+.t5{4q0X3+2<ڶD }&v e?cPFpdd׶wY\2lpG/H>vAn1?3 37Vq/]^\8-E*`S8"Z-qk!`UDF5:E+~MG"\Jl O1jve^jFቻSnBK<{qٍez]I W\Skn8UJ6{b5beH2s0?zQ{|M9Oa(P`MS3hd(h D*(C#ĜcsIÃ'C//F*0FX@d GwZI%XD0`A L":X{ w =QѩE[mLF&Ĝ35|&Ð$pe4QUaکXSSykԊ`bjg= (Oӊ[8L A>j5C!5kHU"õ||%f$/93 Ŋ |t+.w}"m"rѧ<`atָC&wƅo/̤g%Â;?Lڭ6aK, ;;{ˋbQF1#Uѻj} H"F.6\qCP*&?t4txoӇ?_7oWତ3ȥ0{7o^veTm4}uk ;P%-*)1^NiPETWYU`> G&.|4?=Y zDVR%h7 VaqW(:IwFahR_=].f~,ocQ:{XoX-׿ަ\o߼}70| {Uݪ..Pl/ -" 椩b5H3g CW49 Rskn@~rЛVj L:jrք@WrC@Enޤ3O'ǭ^It`nsC< T /Vcg&NW$8lc$gzǑWqQW,ْ{L3P,ː_Kv'؉Cl DVV[ғ$tFw=ex;`ǐg&\Hv%_Û((? ?bJPpŨ6,"tX8 P79 \z=*6#/nMy7=t!/Q 0$TB l K C؃#e1m7d—I=RBr5TqC]5&sYk{r eHb޳?(*Iq$ȉOgkטXNyg[Ed.U=yָU\~kT^KdF$yuzfbkX_Yʜ'UN.Wl зBKBIORHV],f`ʉSVqok7py;ua8S(B;gb^a8Փ$Os]vS,6We,]0G ToTtmW,y1f)6S깾6e1޵<9F{}+wfFgB=yO^.rKk7r8HPcT_E@86*=[l֌y~B @n~ؿ 踫nav6`/3a\ T]_`͂/+ʀ;x/w)w|g nKmu&2kRB) f3! O)|#BrB;8 $B_p,{w2Oծ!<=y&cCs;bw̧2[֛6R0MGiSX<+LRҞJ6[ʆ@Ʈ>kQ̭l>#.otZgEݎg']<3]'я[N\. Gf}."{YfM{I+?cOOgqr<vEsZ^j%5Rd7;RzkpבJH9YOs_^ @W&0E^&fEC[? kmMoy* NC@;[awVoř^Ƞwis@7M)MHף…'lrXm0WQa+jAϷmYyVvت6!$mV (J$}f$!>6XKd0e65Y.ՍOn`FRvzZ2iv-kX ~N,oڋriڳ2. z&;ܪnh݈;ew4 UdѶ4T =]Yw[a~&T fI0'Md[V~qwN(coPAG) o۰wlJXQoYaA>u89E~qd\ը=Nׁbvvcj=A؄a Bh 6|6L(jN~tX^jXcX1[^.+%nwP`*_)_?}j_ωT^jv7kp7ȽY܉!hsb֧Kuy4A'LG<<6{wP 3p'ǧ}y}__꭬^ nTĕ~D^0/k._uZ^??eUҤG^'gk+;H. .zD$>?|QB|j_IX? jgzոG(H O>?Ҝ9 bⷁɧ qg`,ÇS}ig"~\-V?ZBudVt/?<6n&Gr)!e*{ǎ+?6sy&S2@ׇg3&  !1|X0Ziq9MYxŵ{1?0\,'?!d0J6,BVB|*zB+!+^V 9O$i Y?@իhh>ƼH%4fDqV¾xu^] !K([#MOcE RRJ",jRERPGSrb^-wUN_s3?Xj'tlqի l>vEUВ?|Ac {@[bJu|i'T^voW)cl׻Y];q{u;! \/^k>jPp(t;lAU*hdxwy|鍿BLw0 }.`vOs@ mX<:ݦ8Jѝ4 J"inѷa 苌nPO[8tcrN|>I|W v@f+dSz?nUt]zƂ2aSԈhOðI)@l 4gzEsNϘOGLhe9",0IV,6=Dm7fDa}vb'U4Nߞ8KμoMr;f'M.cny,;A Enp1O03֒⹟-~mw'RyVVr>g\ 0LqW &x+L:,TQ8`~J[lzg̤YDN,5axߍDм Ed]T:-AJtG7 @uaQs"RGsSykb˶n% 3os,bC9kWcp>8P^%3B?w}/ `x<${XیZ=ؖ,ѨSkޭ!{ C;fT؅̋8?LOG0#7$*!(Z =GG bI |&jWM\4Pk[m*;SCJW{ uMߨAs8ʳuz>8l#9ؾ<9^5&yR1Ň'cN;!7ãcۇzj-ARFkRŪ SaΙbA#ȁ>ڿuνN @uFzZhL>'6](l ?JuŤU/O MQ1|g+쟩|M6=Aa5khTC#F]ϵMQ89/llXE IRZDx b-^7ysuvζNgMD>F@x,@q1zs cbI!1o/7y×o0Pu6wuzwcu=0ə1smlf# K gO󑾴1>6)Cv{d"^\*y7}x R7mey~.r ]]Üsòݧ>Eof/Ȑ-S^D$r!K([GG;VQpG{]/l_iB|ـPdzyn;qDtc_C؏k2tvVSf.;{G(7VڍRjOoL7ٳ24C!O@ 7"9$n8IN\m߄nʉ1=x=7cBJۅhT2rYA,0v6ݠ0s+S/g+tKEIuU7mٹD_o(JSmꓥH?P>GVʋe[5Lc_'˹j$:䷓}8mk.7gU'( 97lSry7 ^QyAJ{!p@d4B, *B.!ha)^Z j\NXg&)~XP&$e3?1_s0!Wr|ccwQ#ux@C@6C" mIYs†UiXzTHL+ ([zr`U:~-**DRXcH8CQU#z'ԊdDyC+&%YP,("R_R-kM5ִ*5m5&nk>VK,K@<:ȽTdfD$`ʿm"Z`ok3av:(hRdPf7A3 j{-۫;Z$+W>~&}EtӣɔtU/0jj."pfH=q>K)94#%o3QT>I<JD] V|xDچڔeK) hr #g\_0|]q9(Ga 629ŧrUSg%T]@uzq yhJ:R#hl(8yv¤]:C!DۍreE1wdL8D2cMam":I[J(X@xXi#}ܸU p] ulz]8/b\.F|p]^|BT-~1W"3w(3bYU=q'B{Uv]MQY@ d<6JdMQ<,T!6|FnekUV}+`C潦i1:>&,*`}j(]%XcQj =Ƣ8mI6d:rHnM=??x#8q0ay$BDq-2;_jd*xOiL'#VͰ3 ļΓ|v6 ;+V>Z*vPA<$Y7@)}'<=^ͥsoI6d6/Ppv;nSϋHon:L],HZ)̱x˩j5KN  +m~^߯6ӧ77쌓-ΟٲM,n_ͥov{Z^<,#j Onj:_y /oPXY-Y,Fey˥}GNv#Y ; <Y`Ksߤr6?cWPIHhrIEMO2azQbs\鱭v%.VW[|/'mO̘m-.-[!K{gDc޽RnY{R$Ň^Y1:+Jj(#:~bo+㬌s}*wu<Wev?6T ,o9L

7c+V݋f> q)guYxnKtڔ8`[C1a{x 89iק牂ddl+:25fޱBbU %*(N8(0M@O?U)?$  ~NUQ -bD-tSs]}ɗz vu pM7I㚜rl:um.7S?kuϞ괫:hn,檫65ii8pP+ni~"Ykra&MI\ SGeĉu8s~nYi2b ~ 3LKZi2*4vx#5;-z??_Cc[<-%m2]2_=Y6UNK0.>EӝѰ-bKɬ.  HӞk36i!L*?SeZL.K 5džg;-v@=]fG.K찣=ggN˅phAȢ.KY^S6̛Dt`uanhIM ܾ_KgDlIN+k_NZA [8_īvF QFMY iu>bVp)?0&(l{6WьKY22̇+u?B+~PxPźX剩B^ML;j*FԽԦasm'lAf&lCMVYxovD{ =ٿGڥI}Ƥ%!>[JVhN_+xfC|wKRN /LSZjTZ>TP~$U~f,JZ:׮+\`gQb|3(*W N~bQM[lnN'G[m.y|uǭlc7"THF/MJB|bN'auQ<3D[]JL1$o6ꦺ$@WNFrs&]ѳ vJPqر:Six[C@:2I50HM#a:\MrH3A;Q& pA,O߹8oõye>-q勇oKg/6dEOpPg,*bG]ۤ×ܐ%OArRǼi4*& czՀT;&b DX V?=Ƿ<{3^㪞tklK5bJm1>BVH.!!Go{6g!<GZ8߾oUzWUm#w R~tná:h`WzLW892'c0.5-+Lb[ LB%@8EJjg?v=`!"siu3[XAt۩& S jQ a.:4<"qxIKaڛq3dn1ۡqw#+Yy[y3 !M$9 A }ْx#i5[dVQ%x>M m?XŝnUݵa'sw$msa `S_0CO^|B)lE0BMx鸢OFiX-uVsk_i$} 3(Pz8ZI$o Fh G}>rv&70.L#9rd6{˭Cճ'QKv~TJ} ZSX(S"X=-3wϾh{4H ႚC`<{eG=gә;e-ǹ0);adh]ezI}^#m|y5]FqA.1ַl'm@gFjIgi.C xv4TlF'hq7M椪`6'y}]ܵmu[mVZ]LCncĸ eQ5-/Y'zl-LϜ2N  /䍜 2~s0{=rb{_tX[휷`{|JxK:2~yM&v#4׉/e6ivfiKη ^,n􆧔isEZPyӝf?em?=Y;zW3$7;- f3Ԗ"d_C[PGRj89b˅њJ۝@f>'vaT]K(k"pf[Xܶ{Zi`jkv ia퀜.ْki NdS.Ͷq{pS"otIuS7uei`SWisࠦCK!w1jkAM6åd#c5E9k:;2Z[y]{Ifn:گ= :w("w'ZNW(j [CiE%a}҃0遃\R82FC%wlioi,* Ҭ4CQ,[5տS%zo,rc|2!-(2t''0OӛM# 3҅.Ԍtf⌔5{d2`33ƈ¦,dN"wld(dSS}]w`}\U Bh` bLi#PO]cѩN={w^_;jɲ}1u c]|ڮE31ͳs8?c_w+;?SS |tl9ɕa Ω5&ZSu&0x־pipW _m=1:hMaW.S_$ <՗Q<&O8ڏ5d(fék1OӝŔ21! $c/XkKHRFk9tZ?0w4 %P (u<]6Zƻn04]KKz27{Rp-30/d$ _88kՅ/_Ʒp~!&kh ^h'_G7tqyLKf`b׮Et:ͯ p|@/5Xxy2sb9~3d4ڟ0z5~d,Y,ѩD gkޯb z$Z< ӿfU젧@tqO1Oثpjn,N+ELdOLHe}ozNF7RZRB\3COoGGFH]nSКN ?D& x=9̝˒suʿNX'.* *ü-lnsź5pBYdƖ NG%w5xHZ+r,mpU]\)}ƻgOjowj9w( 3ԝH"DbGb#a$Yr-)錹RvWPG-[Qi4WAt }*ٿL|4?zYU lLUA=x|6թb!pp牘 Ξk{;{pbrE) W6Tl _t?{Wոݳ-K![8]ao.#K2$N]JHҷĖGh42惠=&r t# ˷Y)zg=峞YO|S>)zg=S>mY#!cOܖcy8nRܲxs,UT&TrѫvBq=mFYdxom'Y_^^oӗN맻Ǫq]Y;(;k9ll(B-X/BDnB~^%TԘImBIMF݇i7qU\kp _ *;<,瘶#<,Mz% ?vч΁,y{I"^9OP羲;i}ǵ̟?==U//;8kyoS#ٳЍNL&k>&ݥg_+6AF_Owsq?=ѦWMǛ&؁P,2!]A-".%$0q`qxa@m`CidXԄ#aR5г^|Z -3q")hdq/lӍ&MA}Kcad!- RaԈlx984 ѼgI;]gH~;3Xk1ۧo;91_wϛ;Ve^\|[]߁M|~|94vs,ںB&Mai F-&'wͅ r D =;vyL;oȁHKbNmغpi#E]Α2$9<ܭ{T _E]ZWV ꄲT4ԣ+;h ٸI8Dq5J04UåNmM7тbfQQ]h=XRK}6X#.$^5'O޸65#L8\O9D~ R'O3}:<<쳑/XrI7%sĔhd?o?m4vONcU9Kָ Ĵ%ھGlh4I>H>^*Ji%'9((Hߜe@)JS$+lQ3«ONqhd",t3gg_{&>d-# & _1_qF| ö59CMͅVsON," 2b)(Gx?rAjQF1+toboژeb|'=";ZKW(uB~?I mb0 qB Bӝ\0d^=y8YEF[AҀE|.q0_dJ>b)WŸSCZJR.˼;"LaHaI Gdnlə{$JPMnZ\bc0&w0 Wf/fof}zK[պUUlvyǭZ⚶ok.Prt})R#?۴1=c6A]KmG0zB 6BGXn$!67)]d-qwdm3=7L?qФLV\=iEJ䟓qǍSuuesC~0܀pccç@:z"n6姦2ѐrgKYAT22ȭdlpyOH*'\:MƧ92_| 0yCUl%;+ M%ݡNRVMJZDV -ڹmIR!FMAy`o(46 <b_ˤUr/WK$!֌fH0$r7mӦw|7Eug@M#.f} T>p?4fUr`DX,bC^#ă!(tJٮyg|ݷ?"ȏ,u,&ڝucB (a# =8'L͈cxp|>YWQi=4jhZh။=/m,(6rK$cf;rX 9!! J>$Ez=*\*)ÔPu]`2Q۸r):Qwp=[gdp&Q3A9$XجpBK,VY&:a7D;0{_߅ִA" |1]4]~ٹiv/UBn??Aۄ!҆l}WG!ɥ.~%׷%"a@)լ-WKJq>R:靡wK#]]ʕA:\yCvV~awMa3v藸㋖ e;(U!^hn[tmg3k5ufUcά2jF nl<3IUd<鶯 Ag::9/gZM<%g:]~Y_Y&I"A+RjH "`Qy`{M,HoG/e7$WǏf`Kpx 4HCU=`m!:"3}wʡY9dYgpe%pj7@h`FQbC#/wqXNBqW^04`Õ CTjjn4:ǐ1GGS$=;Qn9JgYKNhEѷyg }

%L%cMEJOHR-?~J烎CZY=tVI{GR)H1I T2L:xII $!>/Z "!s*"|MˎU=m&b?Bw`l[[1a 8r Q6+3N}$"\[Ѭ㝞}WpAzqOH-??ч&a}E>`ɁG$=47#6NWq]so^[ >607o!`1@qx|NmBXo/7Kmntv~BϯM͛p OQ^Ji%PH I]q8* DF cHPq`xa;lesl(V9GV2'(>^5J5NWP(^EnF'Uޏ ֦@.y˪P6<.WCzgq1&P˞59tp4'!]L%(EVjP&)?]i#D*:)Wj @XzMT赓+ \ѮPdQzu<&kd@oJ2!SlG<@#m*WkdS7];kBi2h)ATux%/G;ߠL24?d-ɔԱKvS{~}Ýw7_l%7^V=.L+OD?^m#eu5d_y~TߟTT\}J\Ul`_L1B>m C0zBR~5)㊌ggՇj!omVRlSǫ PX*5=CF!PU3ɲ? t(#C_V%WBz[u~\ͧBEɠ/VE{\zs 8Ko TWdj PX0T{k@ o\Xju2Rjgn "KJSQ)a[MAWcAa (F5wn= 8lqp"YgHnU;@@Dv,5>/ba!v.<|p[`₥T^sC8ū#R;pK!2KBWm\}0n9w {z/hf)GMslW+7 Vql'|EY|Cݱj>=;OƊX\Ig^&\#WȰ:eFF[2NҪ{I:\uG1.2VPqfd A!* ) Rߝp4P0d2ZRi 2]9S͢nB<3ϙ4q3SDgtGUp%rnt33|r$j9KjD-N`*V7ܚHkf&)9,˧ziv__6"K>f'y(VcQE55νS,fnb,6R?*d]YY ~AU˕?0Q?Sy7)v=΁9 &pJ,x+=3jɜ_r}TPNZ-9 X"cMlK,h7n 7k,tyR鼣E~&B3OiJK ( /502*l?χ=z;V&ɼTk Pa$C0<2Z)'T"1^PYP:,Ar >$Hb1JDDAʥٿ:SQ*,;xet[|e٧hI8'ayWiǡ442ء/rg.wv v&o"lV/pBomvBx{aE}OA<s3+)r-Q)G-PBT+ CtpS*2;Au}W5BZj谿X1cŃb;ٝ/%I6LG9ɤEF[׊Ӭzo{Њ]C6v 0]I᯳H)oP㦫_~lmX:Ts-%jYn:%,\}O5vJ/$5Dy`Plv됩o%c?E^.OmʱN3s}}.RQ5xܠyQMlY<y%?~z iT1́•;kFGOC3b8Nk#))a'ےe6VKZ% 4XqO%xD*`Q `/ lm# rӞS04KȣFVu>t@ "ΈRR֏Mqhpw뤵ߜ.>eW)ivVzV 4|6ۑ! x q)Eq?#~B95&"/%%VJWjb=}UvC|4#vC9Y[֫RYp*CA,7Hg0 F-Rᢙ>dp Qtȥ*B3Ai( FM8Ù#LC`/1FN;MAZ"*C/g=-C[9Ƿ?X_,Cދ /ȼ^e}}]8L:WfI(~lyG r@O" ɔ}jtijXJm0!~n*[7dVcO2^6 -Cà_\]0E?ȎH|Z#G0:R2kt3{tY#?0x@Ɍrcq+˴x%@иn'x&|7Rҟs^҅Eyp#cv' eO_;y묄Gb9v@`PYi,l夒Aa쯧\7b qS"-ł@} Hu N(B1+5^Kt[(Pyj\L#?YWs̸muY[getqʼnXX0hŽB+jZjG&He 1Bʘze:lGOǫݐCL@|"it%k2\N!pHʸ)x4/D+ sHpܝݗuO p"<7^ i3|X9u*fB3WI*xd ,Y|bQv4on45Ό79c.]JB|&c; i6 oǥAڤNH]x+:?Z}n)H{t_>{҆v O<%;36e;{}a[Vy9DRF:oXb*(4>IF_E_x݊`@1oO|/OG;?h gQyFlP9TsS,vr!o!G_R&,V_R˦e+l\Rf)[J *\կ#uJw|]Uġ';Z9(yG_4S"- 00~x 06ٜ3'xS"Ec"ܽ%^Qٽ1 SCGeOsƿ;{p5%F"36F!J=wJDC:Vpa0-׿";֊R(#e= oklQ/&))9#:Tc&߈ي_;q֊vYwf7(O;ܻYTz t8׷ۇr W>oSK @txA8J&\P; jGh!ra/^C,#){#j,aOs[j_Гd<_dP%Y$8Ð Nrpa95:!E(UTyx>@+d2` r kwaq S2E4BYK<{,-62a A\][.c) F%H.q8c%s$F8 Z jpZfN˜9픜6h;S Kתm,U9N}Pln^LjS Y>5 <(-%2` ڌJA8o& fO2RѶFBjTPy4 c7krJG6H3H8&W*<s3+)r" 9m/b;Vw]^\.Żrѻ\.M\.˵2_w]f\.Erѻ\.Erѻ\.Erѻ\.Erѻ\.;].zw].zBȕ\-vCa=`SN),I9dPPQ`Ěyf9}hO]t>|=WO/&Z/>jFO2cnvZlGeѨEDAĨFyTKG4T մ,U7b:dx_UХ Qƍbߍ|ןFGOb9dDYj|;S!ՌÍD @'S'LS8R@54h p U5'& Ǔ EC-Q;3\ @Y -2HI)mƈ . @=@$lfvխCGQT?PUvMvLGc:7_el_BXQsF:(hN\S#8W>+wڌT10ah >2n[CH`ma9Hs0: 7l8dt66ز[R&1[i-rm6ܸMA!hY"%@@LbuhWzRfjXLn=~)zJ ㇇2 cq#{U1SA}" %L\ݛԐ"7;Wq}5̎1JSf|YTaAgvc0I..,S'r"D` *OhCl؞Yd-8,0.5!qTFlcQF&q9h#BR2aKpNA<5cfog/q0ȑԞGi3igyR=j1CZsE>mkФXt>2JbƅuVQЂLj6 )DN,*9IuPa] cA|޶:g}Jo›8hx,˔c=%S2qg6ׁIg ?;WwWrzrTDշl׏ޗa;@$D, Jpws-U47(*6Cь,\*O(b$ق65ù5@~ /.Mv;͚*Ni; a~v ;9(!L 7fZciy|1u\a(ol^ [I0{'ԓ`cp^ln'm#a1YE)ǨE -Rr#5u {jjEh(!b8hMub,@.F i-|H#Ons ?W'̦nB'I$*f$EBWj k0+3\Y!ш'*SFi֛PKلE{A8okYL0"N]U;&1QGW|:{6FD[Qkw{ݡO%8@)c V҆Vh2P*W6 \k9v‹c}Ǔ\籾i<~J ?{;09ZeL Q58s7W&@!SyCs6d:L;ZOo,CHBѲٳּwr%~8~rK9]ݖnyró/n>U)6ˡh?H:mtB oy#EJ\jҟt\)[. ),dak{0 aUQbkS" +LsBЕKUQʣl P2t!E鈮.zi#ͺ=FM.j>u* ilz SPOm&[6frKyHe=We1 &cNQʝzɱlQL f3i6״݆Oqiedw+ &uBJiHPG6 :J!Ʌ$VnO 6jA>G 2BcøOWq*"sEv W[ɱ0!"jYh"w/!םN~2O $d0汆N:(KeFBG!0'3,og=f/{3_{T 6`}G^(JalH;#IF/kmRZL<:*nC#gGݳ 4kq-'.ۭNo$9;:r46][˿ޙk_B@~ "gZ})Py L.^f>E. .LNB_CkFy4Fl2l;/rFI;CIp8,뛟j0m'GVeZaz4A!cԹe[UGJhzh+eqmu^j<?k_~Wq4֊h'96 Y>b⌊uX'jG ”'(vN$#)o/ ??;>9ZE}?fAY=S߮ת{{CHFaE*4Ƃp : _HU!f&J nT蠌N,H0N}w-S~z_Uo'G.zwT5ZWjݝʒ /zy,rZpX^ D`q]&>A>]# pک!X c- [o2`0a`VI D0Μ5$Ñ!Bv_1c~_<\~v~Q0]Wj{) >g8G͔4ń ă߯Uت-iH-i*XWkGre Z1e"2EY^eĄ 5:<%bf_g㌰63~_P?Զ+hEWpc>.UzXUǻW=٩[zvru3J"hޜ`f%ELi*P ab0APR$Vhxe Hyd %2Z' Kcdp KLBX#RHDsZA-CuzJhLDZjJq #}@;G@5hj2Epw?UDFN0ɵ{\ ie,X[X9faX)A}Tk%a1a䷼?@8_S7wީ48Z07(wޏQ7G/Kt!w{iw2ri_kB#ɜ0.Wq0Ye"8$V;:ͬyo+N o;H:醽Fw#?0mA?mώ~LnyKgL#t?$ %g]hVËFHUi79h{ޕ7X^;.4_0೑_>pyoF:qkv|J{rˎ]9r:7<ș.kѰ6ۖ#hr~Kw0iA7|$ԑR!cSQKuwbk)|,F zoa,qaYx{lǫ";@6wMil}Y`ÝRp¸yD=ϣtZJsBY =\JBW-WNW\_ס+6c׳kXC ,ޭ ^GMhNDz{(:_J۽(nn54J't\q [P%[uםݳy"O Gw*zWv}E43hk2_+Uj>F_>LP(%c7 s Mwg4X|Y.,(&i5"ˊ hLuTrvԎⴑRlN &SgK6/.m_bv,wOFZZVϷ^u{.'r>,5QcԞYF:vId}6m{wŴ}@A? g?^X;M~6-!h30(((&!Jw, ~fh8B-眊o=|wYu! KPae%bU<ư&)<"Qb,c̙$b82N6b 4NnڦIs&--hWx_+蝹99V]AwzT`{!0I:wT?i:W~UImvUf9qT73g?{ȍ K~jr;WlKErײ=y9PW.S 4ݍ@XZ X =dMnq o PƗ EaTDqbV#;Bf"a8=e]#j/o{;|wEE9O'H7|d?u>է^|މfW҆3y.l!#s؀Իhw XF}3: ˥bQ4./kt/N>aW滿y?ql(8SPOƈ^^~}¨K:o,P@Z ptʢEY' 20`nf$k*؄PIo ſ~ҹ`^_7o/ B?!&-!A?ƁU:oR3q)$TļQLFDo^s8=&8EHI}zf1.__d>Eb5h/>s v^VGˏB&Wzm mΓJ8K0l:#0YgekK[ǬuZǬuZy:f2rcY>!i%{dzKzA~.}6j⌎vju.9~|0PI/\г1W\v6BFayNU|m좓W0ps;GLgA(})lee____gWٓie]0r|>5}u?[In frU!+sV*XY]f~./y+_f~7dֽ#pME;oWvh}.U6 |}#f/R1W̾3|ue:&B:?W [og3TiUʥgЯJf3P_bێ,JalH{lKF7kR:l6 f_H2d-6E+ j 1{E]).A8YUt-P׷lu(Y~aGт8ޗ^fe%\ :>Ym_{ \7/W(]b9z]3d 5v cokog;=u`h{>Z*:uPQmqv* x?1N^aKĎ:ET5} d@܃G]Eux.2:å = ist:I0g ui.-ZI 5%#@Aa}5©s@[ !3?Umy0÷Hi,9g\ f(+й,jߌT r \.҄"$N8> vm$i". 0GE(1x{eX0aArÛ=z:S=X.bU1ŝTLqll4}xyX)a$NFeD20PbMBԔ`1x!a2&A |N'r4/ی7{aykuNn2 я 'wo߾1Ob-Xhm5FljD%ބsz[A h[biB*YQix8$5J wjcN`èq'p~--74e,' /1ilH:l"ƅ44 QVpPIj?w&&!vsV{O&9¯   '[73_{T0ܺ vq' q7rkvz<2`vgtO{-fOut
  • GYe޴z,GU%UU "(OnހjՂ)Po9ce[ͬV1[|:\[ ѫ'VV)Y)恡V'dyyb]g&0n#j)i܋ ٶ t(^ҡ='+o-] 7z\-qޠ7cZz?-RGsWTN6o*_a/\츅\tƔ\rt8[q.`YMƣ~qO"F,Jý&vkqm.K~FR@G $[Gpmө݇Pva]ʯq*mO њ'Ov?d#7nOvj 4GsOVqh=|9ռDO>DaRΟ4OwiZʓE ^#臸4 w:谯gIOqX5eYN ^NV yZ.gkޘ}h%~%:Վ8BR2˻g9-H6bRHғH$82nVqsvZN+P$1ď=c~GN}t&pUm-8v܋`ĭ.F \e /cbT{!|b}aTs^e/H괔+¥cBL,6ʨRk nƉկ<19;fr^KXN[C +z`&zPw`j'n-^T7sW--L^5òT;\ܤs0ŒQ\ @Gˆaesɬ$$D*Дaϐ1@h|AAd4M_1Ԅ gx"%,B HAZI%h7^P4@2Aj n_} |~ts&RVZYws"x R)9 Ĝ3 5`|ILCϕDVDӇmǚvQ$ga`xlbp0ءXz,8Ȁ`^GƠLVUlEICX06v4e<@a4g s&Z34e)%*%R+N ,#]EY9;wyH ? ?dy!|,I糤QޅBAґI;K&[J@\F7k1.ȣ8; ( `G#LPF J1]>?t xWTFd%VpFKJ u}a7$x;ΛquLg?t  I%򯢾]]]NG"AH.vRdYzh1 uֵe smQ>~7R0ԧt'/yog!ŗmEpVA$r. nnsR=BDm$vխ̘W m!PkYA 3 \ 3p(&:Gmͭk _Hhy8i +鄀%0_jXcY nA{6ne/|?􃏓h4- ,KVNJ+KVk%k ڥ!9ogջ˷ߥ/W>^a.uyHaF$rU 3;l_~bhY0^94Uls֓d\Q+}N>\~s c}8pSc>>|ه (;IR6]%pj h%ήX vaW \Ś®ZeˮN]1JYԍaW \c;JPٲdWzFu vh4kb&=6Bfk 0PAR(PpkJ67o|]"ffoi&;{ uGzYop}\K@îKwN~ifZ"rدÙ lw]&Kur5la~T[m>v s|EǺҦ;)4Mn'3`fxYK7-%`"{&CDt"RhRR(Ŝ(aHt|K 9_Mn0wYM{nyoiA$bfVLb>Y*7ty w~{9H6hgW,LXxg&cx}Dq2z}ЬsfZ"ʐФArnfmMZGŇ5ANzLQ4MT^tc<Ǔ.ƝzUaC;ij_+eh^<iQy{?qRor]=OkC7R?zB{Hw:}AMc۸!)M \7&]SwJה51ѱo5&]`JoCf0$io, il? ~UBc$m,kL P0vY4a#A29SJxPY10MiRd!Tߍ`@0/M&DÆNo@/7`g]ҭI&;iwpnxo_۟ ]b.v)wd8&;ޫM/}zS1Ua-5(m *>K[Oo49R3X04 y"ᯠo54>2S!`091&3w7̼.l;=Fƃ%<kO'LjO7c 'H @XG M-UysZki;10BB2PPg($^2!YobA% 9Pg HAcіZJk)mh QIs·ǓaHz :]GSGu0Vfڪ}<ܛQܰ ;-E*`ӥvheZmc^S"0y) ^yMmGt+¥&Ly$,ViǬQFYJy>%ckWg)/UgԞ킦:|5~0Ok}QL&Ie ZpŔ) _^vVnW9CM4ڔb[擟u'G; X(F"rvήSwf&xwz:++þ`MH\:$UX$M ]wUOW BX|oN?|ncdJ?&͛죌b@G$^~6jC?=C:0&dTsԙӤugܼc}}at3yp9-Fwk{R3"@Q% BU#iQ9GTC fY>!< V0b~\i]wbrQ Z=j9ɪQ RQtJIs@_C~ȁNrX3O X)zEpqguuo㷏W.߾~~?_]x u~T@ }ae]CxT fYO|qES9pAuSZ:^nRߎvVXc4),h(l`x F<*c]PNiڃZvʏnG:jÉ9g 0wX=Ki/I4$Zp|r-cʕF N1GK6hiG=g?"J 3F4B*t!m ;LΜѦ8H6k] :aC>S,Dz"7ﺾ;O@-Pٽѩ[ŤU瓆.Zg֫k:YյWZfвyYOw>)`A3-Wp0X_v#iΛFs wْ3е0>)j/n(3r|C^Q+oSz ParC\,>Ȓ,8:Kc1F13=ioɕ/IR݇c'k#`<I N1Ei"Ϳj1ܼ:Ӷ3m6mף㕼[aNY+}v[Kmam!8q}:CGi▱,Px ,ef1grHxJaH ]FCL `pyå P{48P=h:FY^1Tk%i*Ԍ*t㞁T]/ٟ^>N#5`CѾf˶ym!;Ml&&)֢t1ޚ m9AO*~'%G;A'R(IUXְmaWFU]@v$ҢM*Ikؕy^%Az Ld `kUWƝJzMUvaWj˭WH*~*] \~24^P u`J݃]]=t"D+xѷ*X[UVCgW ʅ ]vE]ְ.maW Z]ю]@vE]]%pYkЇήJw+)ƴE &]%pn J2z*A)PǮ^ ht\ë۷GYYՑA:WXp-Pu&|yit뗵 fؙܵAt/fλZ VFwn2,k ){cE0W~ԟf__".e6ZhN*x}J9r_~PB]&Ђ ܼBVz;AOz̔sw޾_/v& a# Y6~3* sw~/;]xAU &/Ti*ʴ;[ /vV4OF*[q"Vq~P3^t 'Y+ LtNɇ%3䝾[#^ XF|0a],$S{+%R <c]\WxDR~hh-JFR>RLSNq飶ěU`̙@sCumI]Kq&Ӕd=H-kynw r{I @ LQk* %^FPLґ0хTM۳}FY1)f5;$ rd-!<0G%:Hc[9u4hL+H!f"pFBb%#1)DXҠu0I̥|zMXhGiuOJk5(8ڊՖUt_ U`QkDſ1Aѡ7&(EE۸GLǧ9]Rrf ϙ*WFahw:ڟK8e]r)oN+zD{qy ,aM|Nո޷-6nˇWi$@wga)A@j$U!#R ʅێV:d ႉh m!M$i2b簫?|bpp%h\?mWLMrMd' _z K\2noOai oA׻ }zBT\-&8:DD/'*4Уܨ#>:.GtPR2*(1ge]I#ىQkPިG(0G@Fb* +%7aFi9|tfP4T v€+2%"MTR#*#+TT!6Pd4mgK:+π DoBTx HGez $X e`&pƒPZE  !@drv#gz > rPhx-5VCxb1%RGAÆEjANjLr~g'uvҜx>a( 23HD M" uƅXHКfpZ 1f8LW;+u%'uU֟:ǶݦvRր4p +'ڐUIRG@2)0?Csh]?֛tYt-shYnM;9oB7Wly*y:cop^{\sڹ뻅6r R;jOrv`g79BZ6nWL68BcK]ΤB2`H$yl|{4Zf-`Ep0F;"V"b"lc^S"0yЌW[,ap)1)pZFe)A@ uGgMb5Ζ sRLݱå_'/C_wdYia_Q&\:$UX$eA\ sg!,aQE0g59+S0*ʾ5D'˫G9ŀTI$N&qܴE+1bTFdEJ::QqS4Yl}5}gq\k]tr1{fQ9[$k/ {R} Ejo.`0xIڑ_?e0vZ^_;"|Ba0,x`Z$ɇu1{ɆQ"kG%h$F+JGPHy8rWҁ >ٰjXcE zuzWg?y볿?aϛ^$8 u 3;pk}C˦񪡩b[ ]kjֳf\kƽM>\`za\DSr.8KwB7R2g2,vYUϷK_:sv.r% wMPRpCIRp$)oK0HQ奉DyU$1(Ey& 鈵O*w)*3u`+b0&l(2: [=5^(B(XSZbm 'ICwZ׳2\F\ֺn>֧Z3V oYW¶Ik҄IQN)93ZMϘ€xTmݒ6xlJvYHBsCVՆtcR;ĬW逰  &.ɯofPYu ·rn˽}y06Q8hw?6W&>'Netrjt5PmW$V!LmpP$]_z(l-$|g4MOVZ+Nks~>MKlzR9c6([8fARL: l /d)| eLNĻBiRf"WKq.S%ly aY۶+LfvLv3v&4vf< Ȓ*I\_{@ROK-"! {NY+?-I)9YX'X`D$5wz0Nf41CpRd)9&}v$%SDcK)$dZ;KIj64NNsNNiAΧ8 JjESo#6gu:Ɨ[.!R))46|p24Aӌ@dKb`.&0 t#t&u|`SU 7vgcwιTp|O{gZ>%$ HSSUI2"RL oX9$,RƤHV||:]:^?T.n axtu2Lߤ.6WfC/܋՘J^c,6xʌ[NTYRop+nK-2u\apj*R,?"*bq;&QN`SgSeJpRC6&f\XsMc en\'=32]٦q.}5aftvn-XQhl;nwσ17 0| CpN(af;ֿ vPN|d Ge 4||0,w\ m*Sr(-X Z,了"xWRuYr?0 kҶuȸU]8ڏ~H}r3!F*3K-v?UTaڃ۠^_QNĿ{f>nCXz:.T51HI_^=-yOgƭƼUJ^ʋ.zl% q5`ڴS4uX'5ƨ<`u\r 7;cޚ'5F_mh^ޏN5WQ{F{Cd37%{f9B9JS_ Fi mnЎT^g)=B2l&}STF@޵PN:ku]z&M8!%hT¨ADoNß-{1 q>'p _qy,xg>l?:pgt+rFTT剢`:*NSታHI0)TKͻB9m+*z²t F˶"beZKŠFǃy+փNԡ) L$@-(_W{3GMA\>!h^]?'IVEh3^Q܋/bo?sjȁGUSwDEE }߉ vM~ ; ?~ݚ Q‘og3Oߤ69Wp "?=W:yzہ@$=M:M;.ދ$ÜXsl46]~@!rkmMͦ|||j`f]os.ϵ ێ9H\ޖ  T/i4WP }/_QDEotdG/ACA+Ax82;ƒۣ'x m'_FsoWYؐod`M/vwՊnn0bzvat64w?]~َ).eh,{1UdaÑ7տ'v`5℣a$Zs*)W-ݞ9/ܜ@:|{i6-0uǾv{W?9׭87,QP3`jWd/i2]4ymlDdbO *L#4"+ۦѻH#K#wFD8m4"҈-҈zӈ]grfr;K#K#wFD0a4"{J|+M&HFI#Ҥi҈,L#kH#'bI#iDB$u_jDŔ{ӎ鏋,k d6W*2ױ?D>ҁMx`Yz4W|G]~:]s0ŒQ\ RGA+ӄKf@$% M9P $0Ãض]X3qY#P,#Hij!LҌ[$8`/( r^.?۲b,v{fU\9ރy R)9%)Ĝ3 5 |IcƐse4Q!G!6T*Ĝ,-kdSS0C(MJǂ hNK"0R@R[iԊ!o,bj85e<84g s&Z34e #|^9l$[.>&#θ*|+.3*k2B$\""~u?|$IUZ (0C3 1j˫Eդq6Y+4_N/JR=DHjP|ua:#}[ТX8G\4 Y8 -@fX#:2*oӾ":6LAJQ\4ꢹmX+FH|1W'Gݲ+:n/4X,딯#._Ow/~g/Oߝa^ٻ_y5я@ nC˺CSVLY׬+2s ƽ>\)ph3;ҭQ!z{eǽrۅׯPs90y*Ԏf:uEn<呭(TRtQ`Clf, ;a <7ȤGڦ4ۡZ?qGIRpJWrҤD9UJRZ1mHHK D<0yq>cx G u8c)[M[[o6I,NS#*a7&5PdR;kPrPV#*X҂5;l~pے|NJ#.PMLЊbؖ[];[[mSrf )v"hD(ҮmuFٻD$WP>N~]iG~Yj62q_dQ`MaS&gvUDeEV! x\uBURgvyE[4@Y؃U0H=躞w=W}¨5i}{E6Evo뭞}1O>ˀ5hz0فNWmx Ѭ:ZZ[woxpu^=ܯdO˻򞞹[|聎w- bg:bY޼/z=k+\>i۵Bf=۶ifi薯(^[^vtMg;wėc\ھY^6 q=OLY9OXX ENrZ~z[~;2m#.eLkB<.sKsQq ?ygLX `4oOC<|A՚`3g׶^}|;|7,gwmo+kqu X%M<1f =ΧC?Cћi~:#ngwV@ 5}mCOZ*ʇ 'TawUbክn{sdG0Ǧg(^U)ݲh4vim_kfv M|d[n"D QdD5&Em0RHs=@l'F=>XXАȜ ڸ.ub/Ll7c>K٩kNuX!'d_*䀴},ZR!r*>Ix- ~"*Ŀ0Y @,B*daH ΉRT3\ysg7c}Fqv~vڳ?6SJAQN ʮzce.[ T.,/癍2g A9u,HəjH00]6 p5ZN,X[lzC"[`w.u3CuvaNu hjWagӷ ^9~ :C)vsvCOL rbT)1*`%.gS<hXO<6ÙvY5lZ!f <[󌿡^k: oܧe]VX,Y s˳.`WWlX((`tysbjP鶳WNJ )kYKsfۣm|yr9;\(D!E9U5tޢw1V].[g5Nbı qŐ0bP.>U9q}v1'Ϩt6R^ 0LQ&5.pEpkiFPd@j kGVus}\ Qk{!o1-F!5{}&JvTg4hqF{ RJ({̀uK!$s|z?"/4xP"p@9ZS+܊X?F>͂4%7G[3za*-g7Q`z2Ƴ1fW;oNL3TԈGːc"I0US4"ӛ K#.t7 ٪2]+xʒ݅+,7VV鸝"O|%X`Ȗԓ۹BǨJB.pC6<JwZqeN_Bqx?/wL(NM^cPro$Z3-vYGN;nq[O5v$}2&gƀ.03x O'O9 T߱o[ kGm J%ɽSpݜ/ ƳE<.`d J. !ZU'#+y@TZ."P8EKTt&gK``DhÈۣ! f>~a0=)f(rxkZ0vad-XSl&$nq;nɡ{0>zipzn2FS}ǻ%"8;% ȩP|B; s!m<.zRjf uai.7A֢4;|5lwZCD?{䴳FFC4:᭶&H\Ҟ-QR%`mcۅ-bQ,j:St~^Q9{|vBZ;=ֽZ{FY'i ye `u* E@WHt㕲4{]zN58I!PlY+1&AGJ%3+ UZr& :a唗Z)IA|,9#cѸPh@ٔq |3%{gq;xk8HVmX)eqa3FeYw 0,IX==}4/ML}WT;vwoz쵚/ww'UGK?g&k{TIX-f .P޹R&jL L}ĤXngY0Ă]1'Nb"0mb)Z($i)-7srtvfqLQF9͕Ra85T|Iict0UKulg*lw$o`KmL@y,T %-Q(R;$ .UTdN=FbǙ5PwaBa͗Fp:cE`#Hc#' K ōjuF|T'厊b˄C5G2\D1!#R񘔀r˥7ZGSFڮMoSW2>H5gQ0n3M(ĄKcN[%2P9F4IVhЩ 9#knEd{Foo u"5>ӱڰX,VWqZBbDυFģVe Ŝ> ͫ~d),%+$f Y:& qVA_Umh~S mO'ͿF@+mXpVqosńc1ȫaiNTqbPs}T}k5Z /jyw Kx/"qVpsvPj^<mqp?V1"jc`A8>+i3NFN /3˷p+?^'ϮV{- 7{;_+ߋ?f`cw㟲HELHQJ:w y7=v8x∇sL:l+gH`)ɹ35,z:YĘM:lhYl:>JX(}\_ԞG5-bPk50/! #׸q[r :B6AϽDN?ҥ?,CGq,bܛC |[ ֜w#˾уCWzI}<''cm{- C}0ydңo6\EX:lG]zwn+q|:T3'M^E'|QI3V9M)w] /I(CMp_<§ }]18hQ[1,hՏ*D w'ylz|+æCN16y-](MR9)sa0C>>rvˎa$0uhCaXCϭE~bhr}1Y e߾Q RGW\rWO?`Sur L1xઙF ⳿ܬ՗WVIZ=Nyua;3쇳bXxUlľf{~zB}(o\ah/w9/Q-_?ܻO_|> a7u/,!. R"߶ nr`V{kvhE?ۘ7\+\t~?ruuV^;6}wMG);]ɺ} ;x7Ȕp6w>ǷqGU}b/^?<,,?#29לּ1VGe$? <.jGn]ˬL浡-MyUxߖo{W \y.^.9n/因=EV"9U6zZ"2YS &4-]yc߾_oMRx4luUoBNo\s\ V S9U=W#nb}0ZevR Lll3VMףrdb$]".!i-Bm(!ƬR֪j%Uنbs4JcFАLU۔3q"# ǜ+/}fF04Tr˜!jZN Ni-JA*4,Y1'LDE* #{R,CMu{Mm9(X,jLQi˵)SV,`R0yS$"D9 $3*e%d=|cRFG%Q) ,g'@DK3yRZճG,3#Jd6LemФ4IeDPA+ # :SD#V5ʁb$W•á-@O0 #AmX/!,JFc5Yk(c0H>٢?;I A1#G"aC5VzV YJBuP$Zʹ(+JFxc:CݖSKQ[0]Np ZSΡmK28QIRXCmn@cpy2UqܪB{☱HāNql(b39&G..`toXaطULiE5̹(,9UB$\Qs( Q#(ԠR3(zS['lu&UXI,# lڄGZ#I)X(fR_ S`V3(_uLb2ji.> vym$r|F T }+)fCf6 a)  <",DJV$%Υ"@AGEG 7!w@D8% :)ؙ!Gr@(ƍ58UT&/ ɶfdޛKhShD\ QQFEWP`Գ$(h@k^OѐflD)fnֈj{RR}YRL#bqыvUV@!_,k (E4PZMil-@ժ6q1v(+WQmBھuUW ;k#&ζشj뗃u[kw^Ko+f<kcD%Ej$%!`GyfEfΙxaWmSn_k_{+愬G5 }6y2 2^,,4xi۪dpSdPl.hZʭp1-OSf8qc IE ڃcHĄ&5-*J2EE `ZcŁSм$*:v`:5ǣ&P܊`m@EWǓ2YЩ֏DA*3b¶)A";>n ӭZ.m8#d*"v1IWCj Ulr^@g!B zv)9Do ۷?I4 Q'?|/UL`QWS׀R"d߾Zc,C)9i06y-hG;fCda5 *I#HfRh ZA+F2ySNH["E[cWI蓋+v#3(.:Ni P(ڢHMA՚EbX:@ 8W Q:7j#J0^#z pNͦ(pcV̍ƙZicMZ |"&T3),)I :-LD֢\X{j:?*U[}05o&KvG 8ڠziXB~YXIg& 4zʠ}!1?n홐ƢG ^v(t mJ.ΘIR[2qF<`\t9c'ِCLX+ ˥7`ebBul=0 t$8tS8.lIҺ=Q)]KF, I0P5ۭ^m#wuy7.Vao:⟚pL3Nbs[Jpf{4o_`t@$C{SynK&Fa짟^f|bRl6͖)]AN),mG%Y.> Z~r>[?&fw;9Ng\Loӗ/ԃ4lpyL &c9/i\zuYD4c("mGs|kpg/ԁˠH.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.BVCmk1 uau`.u`^CZHم:_P].B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.BVcQ:P*w4BXK7J]% u:P uP uP uP uP uP uP uP uP uP uP uP uP uP uPػ7n%W~]g~v.`wop76zFkMKQfl_o,iQkMɪ*vuRNJI:)Q'%DuRNJI:)Q'%Du6QGa;jwu.Ή:VOJDKLњpuRNJI:)Q'%DuRNJI:)Q'%DuRNJI:)Q'%DuRNJI:)Q'%DuRNJI:)Q'%DuRNJI:)Q'%MבTg#%лu{hƾݿ \IQv@I0Ntut%VCtqEsOt +Ia `c]\; FA@Yz3xthL]`ͺw8NW.b Xňu"Z,Ν"ʴ q5bӇS;%'^-=~(ٙ݃h=fquCt+'CKUDE +¸8@ ĝsDWHWih*ݱ"W8"ZΝ%$CJɧTAwav0 T Iᯡ[8z>*i[AonaY]y3A>yMrh8P1e nsᢪAzhf_6*ן~:Gfhi,7grc1B?A7c~)Oe,wl$'=@d_0n y`,`g!Ynu]#Vu*LOβ% GDBXVG U%=.nj/9vւ1|%gSt<3!li]h"ZzMDqh.Уᜃ!vg8+ttQjJpu܁V cǷHW.]V}i"*߁x*y86~?s()Mtutuug*+t Pd]]$]iRh6;UKxW*ego]E<]};tvzv$g ^+l:(4{?6mb{Ktҡwݡ+ ]ZM4R.xEg*%+t2rtQ6Mtu9tEA ]EvĤ+tR}tQ6b]]]1Lwǡ'Ĵw ۷Wi Wfԯkj<d)BB01Ꮓ׍e䓼QI1 j8&ԇޏ݅P&_7@zxx_equs|a=~>oRzPsԜ6܇jwв0hqVHp~XLUVSpƿq?䇪&@=3[0!_zkc jux3-uoF7JXem3<K pO֋nOo#]?6y 4lyV꾈y[,kY >"{6Kfc-{6Wύ(Yq ~Ao,QͰ`Ӽq dag">'|4/ͫ}߃_x~S_}%~x]IMƸf3BSq,wsZSM~_Y8ABե7xC_ E1qӧ| _Tpͧ'l-E fPxmOxhCtd0n~k][<+=?~Ow|3oQxE~uT5 $чEUSGr72뺦zQ9AoLڏ~t|݋_dm\zlˀq?}UGI,09R0P&^j K0+CʘE{j0?]Y܈ROs:!ƫySNͮSJU} ;C+˧gGY>]XO9<'>9 Ϧ[T}hRJ7y0>oAyd>o}j;-;`ڧL?n2.om4vDk>xR\fa?]kq>daS~d 0 .Kz,KSg2cw ch Ι*Q"!%K9 Q#mM*f}mp7PdbukU Jl&*fDp7=`K YZULzy6zYI3hfR̤S:i9b%xexZJ:,/OLDz٫i'NρkJmlB{limN }]S%;'[YOLNogQ+UZ8 Iefe hydL[:);2|l:i~v(VԹ\x- h9sVȐ\l|fE3AjTLTYWǁAxe]S1YGmX'z5FgP%+Lo--6zz9 GSO?/`MPטkZ Umͬ>Gw9NX{=rXITٻ3Tϡ:7g^Ϫ7,\.&2tW@AdˍqQ sa^C-gh#g.ҳSpޟZ$ ]4 <,C彟 4/7tf 1b1ػcn(. ΢WQu?CZU Uը|ssF?~%7P\2VaUn[.ZFTW-!Z:$Im|x|Sw3Y B8kT@CgI 06xU?R>q|ǍVTQgE:RNq,z̝vX-rl ha[0 ^м@ѷWݏXӖ^4&u_z,}wܴZjelՖ)F Z֍ιF-:w9&Ekvck>fbk9xwG0ɈkAo!P #i:P5Բe*MzG2,>Y{5{ö`鄜3 )$TƂd#"_As,,\ Azaw6Zs_*&sfy ~vܗ-+_-8"O_}|SR9MV u?ce6 3E^/9ó~IqPTLiޡ9X)̷s>Trt?\Tff3~glP9-2͘A6ƒ iCՙEU9"W-"JqTQι:gX,qGkˉ)$yrPuPW-yymk>}gtѹP͇䂮8BRm:_!R`Q#a~ 2Cیa2#ψ00$II=>$|UJ-)Q7 :d1b{hmkaoy -+o xJX6km޶Hqֺmyoh}Ӳuoꞵqg8ecv >2YΑv1Ldw&[;*7gA#~]ݭxvg@S`7 R}gfRe_9yn5Z`(3*h2:\9`p'm2mZ2KF*=묤_->Fr;D ,9'׺N>Sa!Gt1NFb w7unK*,ёj9ݑZJXoC1X*as.Tz(4@G DvcM*2Ϧ^e1U'<OmVv[:7V%z,EW4mɔ|Lmx#qIzqzlF:Γ1W)A@j$U!`9`ںꐴ %%c88/L%BJAY&D4rfyp_!fAzr)yza9xyh1P2S<>ٴD8}# (Wo=_Вz\_Qinnr0ɍqQ(+*>w\:ksʈwTvltaKS (k0qc_H>~FK> 8qbSIy/9v.S>Lٙ^.cTG<,= !YM++yqʽ$7?kR֗0:c, %>QB IZ-hN8L& koeHל۳){n2SEʗ+xh}iYw3:MY*iԷ,L~S,gaOI :1s3lVoH’R" 4bXL YQ0i/ d@cJ ZR$ldl#cߐE`C %QCBu Rc%ᆓ> S~:dlt#iIH]Jړ;%zS_f]sᦨv?zm(]oi[ TGٳ92YKfGg-%+Yhq1G0>y }阄1 ъdD\5ebyng4TFNF8@d 9'2o ы%y96;lލle&H׿+.`~]S6LeӹGD{"c1 MɆ6D&>]F}LS;Sܯ]dUQc, UbIGhd872CI 2=*cEhH|) p"R,Bpm a]@#f>*.t4|[gSeXALJHS)3U`)0D4`67/Ü`mA%Ƈx ؄Z(#À& t mDD,P A"A0VB[%Q <:-]^jG% ~=ƀzI{& *qx `W` |A6 ]6fiiAbZ̮,R$&*$ qh2Ǚ Cb LAZ -#LGa$6RaFhbhM X`UX⯣q)Z'osONXZ@g Zq_q-p]#CPK:w ֜Yș򘎁.XO-y^> Tn\'Sz7c`gŶc``B[g!P%\č˭ ŖiD^+8&JTWlbi졵 T8f2>s6ꖋLk'%m s'sM$;jE{{ȣؑsRsՂ&K7F0$41`8[;*8hOEmWGe:wݻ+`=/ Ȅ~8 5m&<-r5.Jw.}6Lsa7֠wx򍽽/ oo/Q.gn q%ʸxg6\)/κt0xc ^lb&R3N׺"Cw-LͱqY&朷8Z\GOuipK*! XGIIQ9jT"˒2PHpR_Γs[XABhD]g/2CڸBoNfx(ncҜsO3hk_ϡ!:ųT9):{~kGH֦NnnU+6jd'ܙ9k#_ &3_Gz ]%أO' $|ʷʇ5Hv~mtō88cK|~-Ght*%sfݢ -#vMVhG)B֠i5L75٢"'D '~(H wyc G^đM.ǡRn2ׇ֓ӥ IFL h B00hq6 IBP-X h&z Gq߀4e4}Mք濷ly 7h_ن59V&6Y|'tG'ɜ`t)O`ȣ?X$؜`O04i#edj3.*E2VӬqO6i &Ħ}X@I%]?'5kn`l*ŰPw<=i_T7m׺>|P@uҖKDREkv;Tj1<FXP%ØKK?0 zxDGۛc k՚nr P"Y@\FtQg,zqё) %+9}$r#bm?_9z`ܳRk] tv;9Z6V[MrKOYAwr l.oh7s;olV;TZteZQn튐Fc%lO\Aøڮ5fmu.EåC:nK4}rι˳}wXw@~F?^__c_%=~o//A,FuVHe|~~8]}Uڟ4A*?^+;JA Bmٰx:oMX}STG v/f;C4=uct'],-X7e^dw$]_rn`رg|ݘ2!ƴQyVɬdj[SRQhu&1~ޟˌPv[Y,β $H aiokZDX$( ^Hz)Ό쁃~ՙiEγXو̷^N8̳N_W"B6Nic|vjգp:q[p5X}|fgTGEGVZޑ}pT^ÿgh~2[dS̮^Ԝg/KyrȣʹkDzaTz! |$*M~u'.m^C.eK>0䯇 C_?;͞ag;wz)굩ZxEN-Ϝ)mȾ*#Y^WMm ;PJmtuu"-ɽR9j٩34zQAr&Gbv/-.4^Me''[|Vi]Lo? -MR}z:6] 6`9.\&N9$7zօ8c]]yӼ\?sִ|9V?:3#w|,nIe+]VsyN..&U%@gJCeSӚ2&1E旕YưY `gVW+mֳ^ LݝeGRA?߫ * "/=˟P!LZҜciKU珍'c-c(_a)7&͇l:unD!ϖV V[ȿ{Pѳ⣥{^b=&zD,֣1֖N.qB3~۵e6태_ 349g:63FZ/!?>>߀ ]r3RVnhY߹+#˳g={U{xa9&PSy6~]߾ؿ.ZY PjT;5. 9߾e;l6®oϦa1~_m4zq/=0Vʘ#|e5*]cKAjw8^p3U\?ٚ`g(/KÜ- Ec5X dO m?Q._+ݢPݿ 4;6iVjC~b3gλڲqm[A";,;v1޾#[u5JX(В;X񗡐ՄN"iٳwuRm@M$oDh$gCu'E"{uK@*>6|+18geqqNZ0 n, /WW@qJ Q=]Y }*,ٻ6#U! %'S"L\,W=3|j#ڀeUu=й%*syw4IhlUDw8~L_0@ŜmT7W@nmпmHO~ u̲p#rVX9+G`#rVX9+G`{|omzQZjDsVX9+G`#rVX9+G`#rVX9+G`#M^7yvFm# i /(+/' ȕH\JVVs*9F`i ߕZ/j:zעSW;NaW0GEv |ukqj~v/k,nX{P~緤w_ ArWY>t2UeA7U~^qVj)Z{lN [&>YDZmc^S"$4#a.}bUpQ4)pZFe)A@kYkU7&gyC9m>oQy7zR{vl]W`T:; Q2te/Qk_/~8;\Oy202L` uk|wǺ}2 #ŕ )zF'9(p D*Д1ȜK ^VΆ MfKב "%!@Ra4N+!ʓH&A < ⎕*en8ThwJX9zkasQDV8S>p8 +1LB lDc`H2(M?Y ZGSAbs!1h`#+x@s `N yH D̂ԯ Xk0Z )b(Y}B 94c ל $(VGMfhEK:)8wHpڮ`bWutR\Pko<#_Z$AK"rCi 1$_G3%aJ0̧ƅQe4uK5A+N]/dWwC)]UEIP%zsB`6pyisݯccySث#m=riT *Jw(I*D%A*/M$c"A)3Q0 n#-=(9ꋗykrD\5T6m8;`+b0&l(2: R{kPqPN#.(`byX|Ⱦc~726qol+xໟΫ>Nީ$s @h|TO'P6:vG(ӭ];?&j4nHm)i38ؐ,-]myn24(tXLUpx5vI,IL㵕LKFJNϋb4OyP 7l71y$: vPw!L]RuOa_DNEyZbSh8 nFnw*ع~0n2:^]&^m 79g)6Es}.eMv4AiC*Caj3j2b=6MVHKD+~ypuE7Zd<-:&WZwmUg?elQc\r˦]mBb~Hf>𡴹_ Ns]j׶wabO7]2q5uX%L|a807=[0{j1zu:7ڊvBxaE}ͰS_dnǰ+WrWIR[0?TDc2C j A^n1[WƊ+vƊlwr' Ƕn,ǔ~Qe4k[Va40qnᡧmB@Hh:o (,AP᫏*#Pe,:uq?@)np0sH[]D‰uXTE(zX%@!@ZGK!{(48P=d:FY^1Tk%i rwMFgw^w c}~ZLϋ炾eD^NĚbR70I!V &c.UX^rPc- @ F/pJÃ& ľ CF_o0۟:)?.r+;"~VU A<[kvie \U,ZXJL!0"4\RM*/%,Yydcn"")I^nȦ Ŝ(aH3'™%c;i≿YU/ӲXۉY؎jz=7MC=>^O|zJum69`{kyqn![dE߼QS.߸q6;Z۔;Jϫ?eŞH|pHgѼ0#8ì`1%7iQh 5(O+Lg#vr\}\TG EmёaYqGkZRJ!YA=UAevT[ms78{hB ݷh(>dӁt~<.+%T1mGp>Y m0bD@8Io|=r!*et|%xGȸ>Yݱ=Fk{ŐdymnӒtTOaMƖؒQ7[ w>=}d2FֲXVXDS3̸{TJ|1âg,oO #z!Uxd!R&R/5eDDL ` Xy$RD{;^Nh~= |N>zlz ̜/]fw59y-Xhm 4wFm6*GT KIwC7,Q$*%mV.{fJHݮcd%Y Ly`|o{_9kK8/5vq! 5-肰)+@@l{UHwMb8^.UmQVE4#UU7 zUeS?H7*F۴WRğ|Tg0[z;a=e*y*ܬ mw<7~>aQ,ṰݞJ)v^'s,MWlʬg{]\!KopPnrkID ZNQ*V:Y._C.K%A2`f|;Ѭ VWa(^>ZXDZCDu 0݉cK2]7z< ]vxW}˥?6T)>Ê_3ܷ[Unr3]2^DPIraag{nuWAtm|7{L+MԼ"#gq\Xvq4 6DD$碯ͬkWi6\ޞ '=F߁\p<\|Sޯ岗zu*'w`n%곜]ț]FB^GA<t5>} Ɍ5PV! $- 9`{ V|D2~o鬣!c&8c*h%{gqoR0#4'(xAc?.g -Ksc7d#mNqB6Rr/&w"_HD瞻?QQsc0䋊[OU `M*5`4 e.81vTJqZ7rPI LZ}|OSZuճ ͆pF:JrIp*IQBd {iHRjmx'*HZWIpp^T8ZA۠FH"I`ci9pJuqޘjnCt m栊juKMuq251j"t 8S.3O=o|5LUԏOTOwt(ҡhyOr^ NT4iGQ<**>:.Gt]ňXp$wO(I0E]{ E덊EH7QL%{d9̨0 <'6!gEC%DhVLZ"$"MTR#*#+TT!Q-lhڳFNG9+w fK/=|_At4Z&w@ lKa NxR+3! <1`JE4QwاߘA8_hRc:'3: .%6jb',bGIDD<#yiq=\Xd!J#I/;À,aO+IebF_ERTR"샤*"2r鍖.e,KL4ZXn:.ElīT" Y fNbŒ1r-Mex.7tR$+4Tk*yD`lr4z[ >S|!,,3K4`( USǸIyc-` s!M1`HMaf*^8Q1HTc411 s[ <pF%R"a'KR;gƸcU8rr#yM RO8iE`Zs"N)} 9Tc5@+p8?ލ]8hu.B4mB;28Ss&2Ŏ:#/"%\}U !9s 8l :%nU Hf &|X=TbnlbM'}g ;O<5i803U(fQF;Ƀ&k(7|L+]b X=:P;BXRY 1(S5{)D%T{ @ IDGeGTU#M0A3bTi@ryͣ(@g)W*=uX"""Hh{d*7KAg]}z4d.\o3Xjh-k /aD+%xibB;83@,VGEu4nQ;~}jTW}p?8-/rS'v_zuh^U[rIN;`V) w^%TmTSs җBht?~#ݘg3|>asl є_] j䋡O4; '_ʝ"_i WU C`Qi;UñkEs~Q}tZ+Svl A(czfl 8 Iē\x +#Fku9GT"]Stq;B.&(;]^;ӕ E 6:fsTn[2A“p<,?@q?=0Tu:ټN{Ize> 722Ⱦ݇nV6vi%MObr(QrGyIw4p{wк|-8Pmш@3-%|SOBH0$ )$fqՈ{FV KP[egw97継1ma5*d׳k*Tޖ<^_R7D$52B1ȸo5EThR /%tQغUDXl4>Ӌs+ka_^wգVתwLm%Ldi0' PNFlAX50 RMjїu!>vRYAѠh0kl q0u^{OCoM"Kݕ\Mp.w\|}Tďwu-nEW႞asZVqSqs9JYW>Q>)ox/:B3ݷiw΢ .~e7.]o:fٲ՗7V+1T1!˫C^%NV,Z6…y1Bڶ1^y 8&#u;quۉ۶ˎbW.xAPʢ:YԤ8'|NrkJ\R?<?{mP9=Sr !B.a)_Y` <gE1T%Yqlr\1$}$LT ç7g=.aX :U}v&Yjkasx8Dnc--c, tiHnǖ&m^> yiVtˇqUtØLWiS9t;)Tb}&H%s4 Jtq{^J4R(ךZ VT>d)ݨJrxڴ~[}MjuJFºtux35'c<+seQuwW$8%i,52HL\L9Azǣܑē\yMJgV\cL-0[8`o_FoTא59iBNYS�c kү{3dŒbTq[80`4|%R>rz (Em>3'sw-VYMƏcCq] qrQtcMur]k]N.UIjj5PZYsv|o.o^]H>Tt;>oko@F JšјAńwqfr\XP6TZiR 8?w=3}*7>UE \b/Q!ϫ4P/W咮{;h`V6'8UT蝧$vL55;LMAӬ׶ڐ7!O&jg6^'YV!]7WS~<{>p08ăޓy 3@ z!|&\)ƽ?v?"<8'0s>=Nk% .k+`ۉ9B5Et}IU ]f]@H7֞ĠwV Fj8qzn;~FwOInG]Jq2t--~G#`(.蝵ctᨤY) 1Cby̌Tiɍ X9Vs}6%'ud,3J[+(F|w7~H^C;ڽŵkގӾV(Ҿ~w~{:e^xwq;/}jޞ~W̌ϧz^BїoFs.&,T'k<$I3 K!hVʔ5@L@FL*r!wGc7C0-/$c RSL&F.,v™r2'Zw2EH4WJiP2Iict [ pYof;5w|y,Yr\SZX")Q(cNb0REIEսBe^]XXQ$_Ih1yhp yX$.%7VTUjINZK䎊be¡ё"\D1!#R񘔀Kou(ptuxβM-Udb%oKu\x{qt]ߋ(~|fQ(9 3Ƙ7_XTh@t@/SsIi SArGh+ݏ#BPo b"r,La|X,ь10WM&卵(%3̅ 6d #6G(C~tx.(&!GChX|i;Fn.A%v΀ƆH0y@B Tu,"B鷫86.i3gr}YTDY<@g1B`v4a ܒuVcͤ@Fab,a#C`1ĩ 3oDOy}ퟲոIU3ֲdž;$`D\>V'r`OM?,{nMCwN);׌tsv*Fj&Կㄹsp M8a)ua'NX˝gJ]߈}mW:jp F:"A^1GK߂B.DZQ&֜H'S k_B;XMv<,tt0g:}yZwצ!65:e89wߩ;bGK`.`ZsX  0>1T!GH :%nU Hf &|X=Tu #6NgY+*gmdY~؅qm8,?8߿ ,f{/`YF<{.g̦O~^f?=ϔXqܗMUZA,|r\Ѝ@11+;T70a{.B(,WM1S,-1J)Id-! JP+<(n΂4/13)u\ eVr!!a =嘮<)AeܵkrAcSG=&)bOt?몖fQ*~_xv׊|];W(ywqg[GAן<2 PHam`8_?L:Or~8ZVmSu" 秠,gJ{,ꙗ Pݛ~s6?6ڷ7htP8[gS75 YܿۦZw7^\wGjގi^vg:M#Ϣ;.y'B}ݖQ?@reLIܕx=u>v0ܟ!^"8sJ)m: 4Q`GscS` sn=:6]JjhH*}7 rT:ʉ~oqT$UE&S>I^vi]dwaxXP1wAQӐ;zx^s#a( K8,OYER2θ9&‘n.Y< |昮SǤ˻"A%Gp L0 jd9eD1׵H. &Gm{i@J!(Dz\;)YPz[D "֯=*9Jz mJw״m*f `D2e{g@Y-D}Q~p_Oe$ʃ2p-P\[p & B/!UJu]T Ii2IJnj};ԱL/ #as0۾gva["I0P~hwَE 3IЋa\v^|ST'!ob* +!1d!d->튀B 5E0& F0Q6e&5-Z?q2-YYZA_Ϻ[u]s`59`V$u* cmWRS72}XM1bjw*`v>$vV_3'I1K0+SўDrY-2,\uyZeYu-P]CK՟=QU)9="cc*j:9MxuMb=- Fq;ږQ3>9s;Q!#o`r)֓L7Ԛ ڍ̩݌"bډQ08(ɓv_Nt63n[g1wjk9iN+膞75ÓUvlKRM2_ en$?jcQoi7DƝS@ׇ777Սcykly 3Yq@` e߯m?kw^]Ceז5gu1י琧{|lgڠ@j#ݻXmq&10_i#"#x&4[̈h$,?[xqF/wנr!09tj}R͹oƵG4e㆒k܌ޗ,/tEļE,#rt@'+Y0/v|ǡyjr(4H8ܳGJH¢Y҂CkBBOT|/vrr '%9ʫN nЫ.F7q3N:"K; @K&:Q{I[]s8*gv҆0;ieB'k\>3  ͓*RN"SO's2,T6KntN3~: cNYqEM6q mBN GJ4kA&i2IB"eJ;yYb|QkY:ݐb?|K v3t@vH+/Q6$4u hfyj'Mc$,/~qk}e 5 1|F@IܝieA.Vv+[ Vَtc2ydpI}#߈,p7*|ʏK`еMܢ671a+qaϰ&u!#d} 5!ԩ[BSP5Զ(jT3oQДop30,.< ? Dd 2fK($5Pa\a㳾 7yT[MO\FGmEGO9ɮi]h0 Xp9Tf Xh/MLõ=]{kQ6]/FwֳgG)6ݎɯ:+_֚"nኖ=K|F/G֋GQV+G xE^Wi'95?L(RVݸO3| ؗٗ.m㏀`[xj ig+R|'m>\n\(<<{W~̇QP&KJȕ&2~[hqiOztD@4u `U?}a%'@P]鬦?Fˮj]߻mLX3m`W7H>D&ɇ$mk[Ɠ?_"4 z*M嚧jAW#+,b4#=oF굸}<6^ܦcsΤFd:`8,4@CȦSyAGp7ȜǗ`86+C&hfУYwo|K ϶̤uY>z/TH#N#P\mɨ4(~5r+302PɦH1VIv.NVGQҖx?GC~tq#98' :X'NG")Ud w9OG}{%urtf/Tg5Tf׶kN~ M#BACgi%PS)g0.Gp"2=47١ZvBw[ЅONDTVl0>!F22΍}?@!˫˵@nn\-}*Ad\z] <!?!`7|&g`JwF?RUv":`߾bD>~Y"Pz̕mGGEhsJKG+WKf`gU?"BlJwUzzA>y7vtPZ?PS-޵> 'BZv` ,щ1|_ T`3_4PP@T$9\0 lҞ1UihHGld4P̕YFG<xU*N0@:w-F=.8i]-`D1{R' 7vjd>1>Obt)zpK> HT LjL9(~> F[CBPīZ]́J(d|'e:ҎDw] lU70/ VU96˗9| UY|CXww7(dE}4FycxI_!0ऍԧ {-eΨW3+A3iτ xJZnOQu<=yd-<_v5%t aK8iX oml~~ו-] kjXuܓ~p+?`c*r ܜj^ * Db'CA j۞ozzvYO]bZ(!-!{gG%Ny ݪ;0{sܟKx9l5UB~TψpUkc❟PA+DWQTp9 ]E/"J[Dc*DW2trUh}t()";ztY k+CW.eUVNWDr] bH^$:@.^V+P<ca绝%LYttr|!^7Yʬ"83!i:T(a̩&T2 (QTrtr"8dg~"6z7Ã0jőWSSgo)Z)@m~׬vA^ c(CHzb О~F eL^ JЬeZ+ӻkѮJW5aTH%ژJ L/jk?Tǜb$zOZ4 ?^ kmlS)S+,mvAS)JZU&}=XrEQW,S/-W˞xq5= ]R=XĎ:"S+L5¸2tRUhPjHW-Y*!]E2O\ %K+DQ"t+QUDm_{}(n1*nnQ~؟Knf"!`[fF6?zpb,~֬R!֩RQ<ײV1T)b+X N Y@SK[,} *r d~tKLѽz EFR\wbe1ŝv}T1) DhCIaP|F&e@Ey%Kp3 豏uLKI 1`Xp)-XX0lYvY&óhg&r.7[goF烓Qoa @||k`َZ!Ō7~l5NOۙm6dR!afTAaw, ]@GIMVJmq:>4PS <5`.?m46D ZWã[.PH2f4jz, +"4j$4|]ϥI?cwzSWe '|ZX?gQKLq<(˘t3N0WJBI2EV% AD+}R$ s8 ]\"HUoA}gU3<9l?"9FTXxGȏrg9G_״vevz>K+14j8Qc 2K(z~T# Ulo~$O$%I{{NKo X"'HՖ85A`U aTO30.>,q(WKMLy$Qc(,<XcLkm0OpRqWⱊm=T\d&GSȮR}783/ 9'L(edeδt\S+HH3z[(/d&R6/-n:K ?ߛ1,Z470T4:AR"2بhbJDJNa "Y1SwZI%XD0`AP-@2AZ=f {zKq6wX9zjc]FE ř4œ35@FÐ$pe4QxOqakjZƐiSH4rGV,89䱂`Lɝ!m֐2_HVb_kԊ`bjg=ћT1]10{͙@bc%Pa _:`y>B8>L;b}NgϺַ#/ |颶ٛN"'X - 9Y,:ԾMM#m1q\)ʸ~ j^ W[~߁ǸESmj1Y=R*أQ,H_bf0> ^CW'ŵJh2Ьia\Q&Č\:$U$ kQB/?dŧtN_[#}_kW&#BJ ʿ& /nոK(uް;.>#Fqbqhy΋X=v2aS3wkǴyH6pp3(>8t#kE.[DzD"^H!4'-f=]{>h1Qׇӷ0|C}"ֆZ kVZnk*D׌l /ӯ#p1lAq53 J?|9㯇˫.[0L۵9>Y3=_W6e%\DTR ](5sfŅ-ɘ5F'sH{Iyv{Jl %O _(I*NJU^(JIZy&NOt7һãGJK29؊~cUDXq&s5;E۱zC7vijS|DI 6B5^(B(O.(`l jMO Ӊ[dA3;{G<[8B;cɯbJnZ2,Ev<άY@Uk=Я/ܼ2+7{eT/_ / 'X ;ïQ`2aZ_خ~貗rMe{c_\)j5CfHJ4U5b0@FweRqI(YlCXL!XKF&5FT9lhuVez>UϤڍ\ڴ;9ėW]߫/ntZfm%.d^e[<{ykɖv4l jC\X&xIvF}oslQd6Vڜ:[M>X'P0vY4a#A2 <Kۺ[[VTV-O_a(K!˸ Y҃SGP.M>WfPfrN ]7׿,⧶WmI~/\z X M1^d5 >>8yyuwd;1}մO!;}ٞ|[P>T,| )6^-Wl\Ẕ W1Y~v..Mkfl`RSFDD b FрG!eLD1ۮ~8읐4|ta6tpGM1ke-e` 04Y;LjnDh:vF&Z8?q ڽжZf|jomwG;Jy!ϟno^ĪV^GyL-f<3Nvxj~97 d6WHCsKFD§oޜw/iM`"{&CDI iCpCLD6,xGI4 GZ:%xx1'.kJ.ԑRjrjN'Y诔򣛂&B$ 9B #|Na<ouYDs zy1 ^3 w:g`"kKBm ti?>O:+cnxִܲ[JEn qrvpz=Pg71Ⱆyx >2Y gqo߻=(ãA|ϻ}[phU0\Nef6_<=1VYE 62( F刊`) NPh&TZD$8[I7q z5+@Ҳq#ٛIp3~*,&Lh\zlrƅ54` RVD/#[+'gt$/@g4, =Mݚd *A9J4w''IW.w./g߁= F>Yoa1QO6z vevrKͤrK}|gC>E= ye-nZj!0wFI&XpXE)l9!pYl`Y#vaYjDz٫i'Mu)iܱSߣ ͶM踄*F>aj[MW2ʒC%vgx=3_nRY&ud'p>fחN|kb:M';.5,&^'^;%TZ-t=+*2~=5&XL԰"M]sA|`QŽeY43YF4NZ) c4v:Z>U.;oC ' \ޏ댾xn^kvr><ש,w*_ɗrB^BV,zƙ|t{B2cʕ(gH2Is/}{`xW{'w`Xmr ~/x:=V)nRl=N;[B7zf!% 3Bs46RtqK#} y{rӧ-&{RJek<꒟cj93A!n&T%դYP| iP8x3kx@s(3lqc@^= GIs4q̟XP ڠE;M`hJf2UcԒNJ#%y5F BRY^Z1To;"`ںyҴ%CJ'0)S^BJAY&D4rfITŮ>#7g5oL:4ЃqZv2nu/^M6ZěوCmЎaNl R<.UpQIt$3BeuEiRV4%[hq(#=<**>:.GtPRDpƐlޣ!4gGG5Cz"B9i&D6p L>̨0 2'f43P"4 +&-Ȉ&*HΑC*`6Pd4mg5gMl-DP`iHGez Ā$X e`&pƒPZE I S2{ p\TkGnB%Z`Xmy HÆEKrɠLKZnmN񽖖rGIDD#yI]֞y.B%0ÄJVq486j=LeCU9WglZ7\qpБQ[M*Ffd j0D|#~.P{ؚFG]"UЕ ;\ 7nySSOTPE#&j"EZ]"mP;;[Xg]WJnf?T2<\20~`¯ +a!Y3yn=2-឴gioԾ϶h̟ôSbx/7c43C#Mz^'66t<zzA3 XMB[/c^45!CȂN7} MP^a.߸dp%O:r2#z|NN/o'gMO^SU = {./9Ȇ;Wi:Buڻ걺@%&f _*8Bgcy\X׬Tܓ9.X8fcw10xG .@f416% +癿n4@;?+EH 0yTn1XTx$ͼjZNxq.6 돽SЄbC=w;̟M'wӏKMDq3xwsZ1=fx.N~9cVLGЯ~忲iݿlܘ~/|8FQZhզVٷ?˿hp4dʬeYI~_IOXHWЀM6}N6VdSNnbdyFJ2Ea<N#j=Lhvtҟ? V8xZj|Wy>dmzWS}O9O߽SgS=\Op.x{cÜ YI&Ϗ<&O y{V`Hlt:[hOtLcR]6,/RYN's7:B_$oS^>9zl^~}r2KuVN7oNN"%ܼ2?h`39ۼYwP;7Ԏ,eFZρ5 2 4ܼtpHo ^ߚ3pwk9-޶?}ꓛo~z /N?r?_v)5Kn>]Uܔ 6iO{^{P|sZwo_}_}͛3l rNC<[B@ݨ??{|{ixxaZsߗͧƍU1_ɵ ']t^Ygt{0uZ*bU_w FYN?f7 :gۏg@fש'(Pn sM'[ X5؆#cG`)E>c(v"əLuUuuMwW|t&e>|sp)W&;R`1<їF˗97Nӷ{5R\>v`?ȥՀ"n2օ?SHVUĪC9v,ā'RsZBfs% Dɳ0?xM4`ɐt "5zV}|c dt|4|_dr'ܳIJ~]?= ~t"sTP>5AUv wx=h|aFaHp6^+ixq+\W!͘!yWɇ/ʫ,O Qf ϙ(JƧde>`#nC9.m1S0x eMg荋^z ) E..kTw.ۅnhr 8~ Ubz10!.nj_MdO l2>;y,7w=l QKGPMl 8c2F+K]^g~28َJ.%eKu),3 YdhOFƄ̼i ϫʋr]:7}Q=ǧOq!o21ڸ@Y-AQ`m=#JZ2 TWxu%qUZzU$b"ΑT6B˖z FDNQkR]gUwu g>ϐtz>w0HĵIkA-%Vv),Q> !h@HB_ x C`9IĤՔ)5@@$VaџӼx34@LT"Yƍe :1ZrUXtYJhfH! ّ]YDŽ=& ʈȗghtMX#˾] +Gn!xDR;GeMZ+~j98OH&,S70@ ʃJU, ¡ri1:6V&ܪۄEG kѠHԭ*dTUЇ"P4 `:fdRNVq!ˎ&,hR =!nAV#C J"K "$D68 .F v),1 Jhf4><e%9$Va)7wyHrIwQ 7w|D8|+ B !R]o}iAz216\"0t0A$C%I1Τ0"*IR@fm @8BVR H" uzI%R92r\`5p-:լ({ /EzeH_u|  ~*<f\VJybi ]ꍎN)3( >M!*Wcd%婚% /Ҝ"5 %.uȲ) Fޮ 0$Uk?&&Qi\;4JK{ЬvE$2bEr|1: l$hD)bE!L2L2q`|^B܂ .PA*z|-yB$P`l,z'0BP8R{ ZH m:"ct[5 ]p*f8ZUZP|cBNJ %:/)z(x$H&rZȼ,>h.g"%8Lt]Xan@I /!x4d5qnny(G A8Cӝ"YԏH?% _E$  l2 Rp߉va0v_ݖw dHIm|&P:g :Ð!]Rm\^ts@ VwK |iMC`,3x  BiL)]#,܃1i6X<8B vp W^# 8LAG&(X)f @b>'xR<8`i@J)2+8)c\ l: %Ru ` D@lAoPEp]#,30c<Gwϼ Ut"E@k?ivY3LkQ0 Y E"I{ނEh[S #B8yx7,!"#0$c6-j؁+Bn?a:,0V۪2m"1Wfdy"raCST\DkiD .ItwzT둥%C< =a%X_YaY Q!|?YgȊD1QgmT''T+F"w:ktG |P0VhQǰ *dBǐYO+:@.Er ܏@c6ijhI2܃csr-6Zqs=HT %8VY&Jb@PJ$i!p"=\5W {ifݎA33!Q])(%6H!* XK0m@h]X@m,FFX@BAfaUu!{Y|+E\M tY{avtD&C!Hn7.xߓt\DY|1ETWAC AK%!|TTUhFI ],DGs蘘+D"9ǥQ9xJerAmMQ\ 4 7gQ˛|"{WMF)Nv8L&=x>0 EwfViaTO Oӭ'?t: _e6ts/)pi#HLsEQ0Cq5;{O)@k$#8uD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:{K)萈:@~8D!ꀴ{HB+"H"eD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAud! ;$,$CqEq0DV'q:JiAD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:H[NKDtu@\u87X=QG3J $Fx7 Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAuD$ Q:HAus7]SyobfG d.,~1M٧ A4ڹ`gd}nj\Lp5iͪ)\y9ZXt:6^|"ίGK_&:7O e~Yn^DFo4ce& u0̹9XJԏNŪ˒0o٣vHoli^.xlyi[JUMGR .G?}~:; q&IcE+ך}>aV=ڴiR;E[򕊿o|Ϳcw?T% A-uG4ttdgqҽ"P[4*Ӑ8,#-DӤ.P_!pN٣FգF[zR4M}DuQ3"ǗIRWق9%T͍ٟ}mvOc& yfOX2 ,Ymc[L펿Ywt~kEM!-iLVgy4-}Fa|=lƣB7ꇉ{NǗ(NJiX Ĩk60ԣa9٣r]8!Jqh~. N71nsVQz*"I}!|; 1EGRc5WӴLIrι˅5276y2x&zB0jZC_)*e؉|2XCA}4ls5A;N5\pr {-I4G(/TK3 _G/?xҾ9WF?ZۤЮ0PdӋrbGPVBerD[j۩l'+mIKoZZGjKI*Vuni[HÁ-HakAX,#z8<(4dxz*_b5/_Z SDxĮLwbb7_gsQ4gK_դu%};ny>g훟nq#a,NXչudC~(IZRx?@["1T /Mri5׎/q 1QI6>*@>LҬ^ۿN.o>**};6jZ4/GBoZE8mMn- Q/hp}?kK`Sc‹YNv-Y.Zֿ+9lRP}@,oXg-4R0y§^[5l;f2O:}uw0KmgS>VQ4PNcuV/N۬t'PdV}_z@:,k"u.o#%xF033B^pL-9811ggf u8F>Ǭv^2^zuj2;ɒMQi-IKڀ'$5zc/%xhY%jVkqSGa6uՠ.anݾ'R~@nzP]ğt]l"-}R_q0V$&@q}H+ēŏ39L(s1Ơ +9Ӛ;XNJIO05LԸ'5ђC8/#ߪ! m췭~ ϶CQ27y }NzFt7_J|n`> [!LfP)W,2q`~`*'$\P"r+Eșӈ`)fa`6?Vpח|'Q!Q~ˈC ^2^tסCX7:O DUttIQS)ǔ?Ֆe$W[~s'_dKK$hN~j#h Ė%7Fޱ`bpyG9KQ; !F%;% S˂5Nĝ\kHcg}:Qâ{-/O8^tN'jr+v2Cf'ݹwy(u>Á۞ܦmT.M|[찶d\M/GYeh7h7{i7ΏZc7iGY5M+u 8̟߯ɀRukkL Jllvvo5rYn6,4g6]=1mP]g>4'WM}恻_@(+i% ݭ^_ z}r+KmwPjYHJ.&-t_/RW&BO,Acy{}DtTڡXUS2Oҙl͎KuX$_ r70o頮=?5ʠLxSJݪ*87[>46U| &$DC1FmgR\@90UoO '=|R ε$Ą s#Š*{y;°~Drqg:U LgSZf¦3 S,cָ (DK+4#1H^t1@jǖ:vz{=mk~x鎻!SކA#nfPAzP.L .$ 喏G 7A"sm5uy(XftMhسC]6=PC_:SⓍjn 2ĔVBRq48K#A9n(Q\nLi)A#QNyFM7Ny4.jg1)TB8@pSU| s?nL>t h\by8^>}6oEx_ex`tsSfռt]Wq |AOz ̗+J⼢.6-m,.Q&P޹2!,5?GBr{-ܼ`֋Zc)],?32MrE\,*QV9tAvfIibIC((rZ(Ǥ"02KYYhAtcZȉ5g;M/߀Ͷ(jJ}[_:ɨ(l,XvI9B(Y0dջ <~0K#) p:c= ,0G=F(%.U(ejdqIs}Q1bC09Y̑E]DB ctJ%b (ͭBz%Aяؔ5ncܺ ۞E'n G ~qi p{[pO۶GL< zNʍe1 -Q8̉t~ӥҴ R:-Z[+" tכ( YzQ#1% evqҊk-tEpJvj4dݜܭO%4'l455lzw׽Z!]Xp.:20H8l ,U1H*Ŭy+ؚ4 [朙Ww(@nѧYS{ R,.I3AL6%& #KKĞp,, @0RB$^ȴ0ׂBYDGe{pj7dcFSBPc!85T%X 絈2JxeBqyчmXEKCKCK{JK;h {(E}u^_}V2nD[`695;z?j[7<oE ضfUi֑;['+ZӝPh|8SC('ҬfMBbpur :A,77v8[M-z>j4./kiC}[잽fq2WuwY{tl.A2(4wL':MsQƣkqֹ??ho>ٻ6dW.pFKuw<8 pb_60R- )Gx?ë(EJc 4UuW߽B Ih6QRi5f#RJ̷>mi"qೳ_T˰cՉ)7M̚\d՗O4h,aVqXQVi}ŗ~c{5MP{g vizM^PCN3ZNACַ1}c}@x<Ԝ@Une2Y/k .SE5"8=egkeP/8AQ I+"Y%FU'k$! !LIdd Y'F Lx(r%K"CgbΎܚ|ˣѻ8;1\ӹbYJBH&W":@U7@%a:ή$ )E*A!Z,#i ˀ7rd .2izEVDyA! ڧMhESARFvj9*)ze̎#ћ4sѿ=uj?rχybPN6,WK[o@.-e!}7Ue ƃOj]iSmF;ߚnzZo҇F-WxqMDxPRH-'qrR08z2b#񄆾^-UWA1P=i\itR0n r޵ٻ*E4j?m:?_LyZRm:ZC2]%^f?Mk-׋Dӏ-Y5:aI.~lۯOhW߯WPp n04?&W~j/|}~epf2N~ dm_REmu5䣳7hC}Blzf>nz, [VLF Ub4:[ϯVnSKnzwel8i#mAWDYºbT opw1K8vI,B(A-IW35"$ZL^!&„BɾAow&τx!;GOҵf +kVhOexyBG56dRa5VACQ 6DC{ pZ>h./Ɔt4gBmfD$Qv!ͦȢejRRf"{,]ܚB.; YkjO=Ǒg6hc+`( sй`DD  @$u BH%;kڈEޚ9cFL&s+Hgӹ,Yi7CRI"hesILN֐PVF$uAҍ\$=v*˖e4Zԙ'Y^+dD:kMQ=v MOL@-/Le=%=͌(p'n6I -C1 JYE=IY0NugE7|n2=Sqޟ۲FcVKc}2:E@m| 4y@Ox_*]n߿L>5F^s<\^\lovu{cλN-nt;".(YHj[=sќ]JcRsoI'[n@25~ԂZ7ScU N.( 0M$ M`VhSwRAǽت<$&3ՅD`śCtY(e3zI @R>[YYJIn]{_X:}OKtŇ # jx~PW<\}| A6'~2j1]ro'?GF_FuN k?B-o|vfN}Kz|,׍lWw'az7/s6`#q86.gttY3sFk2˥T\ -dx awc+׋!UЖXݣpxSsZJM;Вn;B ހP h.4A^iӮM( o{(2vcoHBi D}STt=(zV`=敎  ZiRY2h$3db.E(:#x!ޣ%P(t[ ڜ2쓾Aq}Rw7L^on\?5]e䵃z$ո_"Cr _uS+玳ᱯKIU4{~7厥 ^\u:=ƨ  WJIR Q{C ds)`B[#WM\۸ +T~q}Z=7c_G狋O??/W`chjx3zڮY 3r&^2ABHj&e|_~e_w/'58+%9)JɀMKͶxT"! Y)"T4*{RwbLS. (1)ҒLOfBCf23/|q.+J1+HiA j2%K4$?%Ђ[6 ]m0lIk0z!RNx)]4!\%cBl]MnXm!Nfؿi3uG 09uhWx/Jv(,!$ h j%b,dTX 1ȢÌ^s;f5? V/gz%sqtjYND2ёVFkx9KS= 5~pJ+dA&P; (9/MI jR^wgMgG=;q[/LǍKs%%DrZlܞQspZH2U SBֽWǩ"SC~b7A$I|TٚCdDVYE "N P,.g== ja'qґpfD9 [Dn}!] a|-46FiPqъptc8mOg2Llx78^9+A-MNITEPk"֖g%dXB쵭ϭ6|f(S*_g"P+ߌ>ˑY#N0~E5R7 cP Sz}B^=S9Svt&!#Վg͓qqu) 73-M{˜kD$4kAT왾n"Vy')øB`28 ؐ K%=]uY#E)8]8t;XA^A{{sy^+F;ȟg(+VS"N,? 6ES?ɚ Wjو5p*0x*vy%Ҳ𿍬T]"ShD6n*x਎ujR;^dѕK2X\TƔ0qr@p:Hi VN^vԮ :k>f=_i7~'74a82<_6uaE!+cTZUm-d|v@_xoK"jևـس}FVu!E439irLD&˾?1EbM..4V6ױ^ǞyBX PɀX'uvJUcB KV:MYN'!U.tlC~״^zMۧA):|x#m+_ŏң|h)Jk7F8Z\*eA#&h41k&}o}a?h~.Xs^>Tg{Op|!9W1-x' IT?k&J8Ggc}&J*Jb6d?WLEqR D؝\CaIRdrz[ z`gk~$t## |>rwjC6?^;t],m]qcԼvXYSݺ7/~T?z28kj$Geao$@jogUOvq|1Αhk9 seb1u%`.fM~=7]>96,Q9*śGm5jZj>HXzs׷%ȯTdG)9! &ǪN1:@;M ?roO~L>}Av`}.5*_<h`O??ah0^34|9s2s>f!(2hG~[+bۯOFGlM–E> yКzMA\OBͯ*kEգJB%B4`nZ66*xh'3&>$\s#Z&1:fƘ 5^$Ryfp8áyX^qxcrsPy{w9;Gqm9ڦV}cd84OdMYOOcOq()A!-B4 ^8:C:p*z>k."&2' \I`,W^ I`5H4tMxǫ(Uʧ'|$}U9CCҷt]kݻ/;$7N2Qэɖ\K]4Ov_.Aj5=lzwZ"h$ݽM;Ϣz^jkn~i$*DjvO.|+noT{_X]bSq D\a' eŖEO(Vsfu/w׫'/'{{ uB{}->tm L|m;0{.u\k'{Dmk}*`1_rgk0eNc!nʩbm rȇ#7@BZ|awƊd4?2; x? Zr0hWce;v $sl\cҿFq<Ky()\E3(QQ@ѻE P?K:(d#1NJJ% ̙CNX`"3F+B)nQNz*66Sihzۖ~@bHL1BkupÌ_"Li2#νFĄN[  k_ž'VEUV**gcߗ? zg~rRuPܷY_hmĔCd9 &:A)hQ;6nnahGoq]|Br8h6h;q2Jp䵦VDC>\kBխ=>.i2>#U: vc!=;mMHRD.2R#-CQ$Ta@:GsaBiNuh!'&}\ń 9Ɩkj!Q)ڄ].,x d9pp!&uܡSPR;ޖTinEMx굝%&Ϲ:Ȑ7!z=S[Vsr-a/Lz{b+fȊ%b cH`?;f_nX(NOEVcPro$Z3-vYGNI7x4V:JjdBzfl 8 \8'ɹ{ZDSʞ 4 "uBnЯs6.*>۞&lC(y[ro> aP,Ljrqҟf} U'#+y@TZ."@85VEe`"D/ +oB-f ww f!WT׆a|t&Q1e;q>/kkp:z#P"p9JZ sw2һw2\{ }ǏNgf:xoݞ/IVo#bf05ŽxCDrY#!Dc[V[\$.iOu(EY탗ZQqTۑ*.-żD1F _RnTeTmU}|`OHbR!# ClAP|twKb4u># ۛ݌Da&}]Y*z~YV͊}T}0]b+:l` z vYyHMnzs9,8ϱCkc`Cd%Y677~?-As]94pءy#mf-h|oruz׹4keĺ1|;\7f)^)CONV&$7ly0ɞm2.tq2)wHRn>H6_M[Ly_UmRŽ;;4[\m61:.On)tխ74̻}tmn[V9"DZ0P.L;B%e':sryfųKT@92U'cޞ'=F?_?v4>Lh!J}{=g *`n% fW7 VZZϪxEϼ{ɿꉝ܉#xa%m1ތ)WCr pNrapv6g0`_NqrvN_|+l1_ڋIw=݌oGS1wNmū9Ymy#űI-{Ts]9ӢV's~oE O{l 0é:2zx}%2'2 #ƘC eMSՋެrx}5K,wT>2dM|O7~ ~x@/!]^Ђ 5)# Y,N)>Q; ԎY-q:(')[&I˗2{|J!e@0E֞ĠwV&q<2m ?CCy&,˰Iʡ3}MTǁ1 6IQ@wJIQ*bgI$( i13S%7zS^j8'iONX4.jg06*!D90W᯷=c6Y!B<+lI~Y"OtFuK_24JZ?~1j{J_w⢣&l{cXdJ ?$I3 h8Zi+S8 e#&ƹE^8x@1I)&&F.zO8RZnd7#E H4WJiP2IilcU-]d36o͚{5~^څgqɂO`R "%-Q(R;$ .UTdL tXe,p^;7a@aʗFp:cE`#@c#' K šzZ~6NZ0_ۈ6˄C#EdY\D$IBCIy^OvӃŸ2D"$eY'^b\I) ؖXɼ*/"##g k,Dɡ550( ~G爍΋0Lh"|KYn&Xe[^$z]`.Cx5Msa !V V^&VEp4\Myb6s[$c2_o$R=3no 0F$Lt7:!S td"3{s0VpGoB_h{gNiR /,PO;&ed(xvxD$+NψJj: [^u˴̆7E a#[+Ydˀ0 #\p%$]UIfXb ;DEռMMϧmNϿͮ.oﯖxd&#QHYu2LwZ D佖Z܊t[1 }'2E[iazڅ?s~X iI(J1,,VJ\tܩr=lx7tAӗu{f YjiÝ{mӖ;2I>}~.5z)e@JF(E |rH%bdF/cLߊtݳ8Vw{S7N7nPp0~T` OP fh/#St$LEt!"'U2`r k7a iU3@ĠL1hL> 'H е>r Þxs2e{C<FHH5pFBb%#11)DXҠ8KFbu>x?9-sZCrYsbGS% `=yq:(9o>HjZ᷵f(+c0_ ΊQK)apn $洟>zCߘ㝌F, VrH# be}0द0+Cʘto'W&\Ά6MdP7~\Fkn*2[V hL U3i_2Zkljn65&̂- $vmמ :N8eeƮdRRD##{_.Q :-E*`ȉ Ǽ: VLA#Da &I8FҎY | \KGL-F}iͥ Ә_g}YWeGsJy#ŕ )zF'Cǁ]H"Q 5,dH`uT1 DܤYShbfɢP ށDJNL`B2ݽJ3K`]2p_*K0U5@<ߝadMx,L%H]鯒D>,79T5/\ZmOF/]K?AW\Xɗ(ʚEݼjk{O惏ld$ 58\]{Kbo$@j ]}6GPHZtn!P;Y&O>ƕL4x4|l7zsp;mY%sT.kԮ6+UN=<@>O8S_HK|q_cU- ԝ F _?}9/_//OZO@1\tM߿k~ {вavhC׵Zu+\rǸOهWuk@~.~y0uT\iBZg: L4mVjVy2+J(/]Uɵl[mU4,;NR(vHR͍`R8/M$c"A)3r(@r$#;;pF0ŕe1 6BW=5^(B(XSZ|˹I9&H㗜}NƵEh *GlT/Uf5fO.cqH%V*dI S%9blN.blhOI՚8AjES?zT!$UrW^)ǝ0'362H>?MfL#`Q-HC-r>QN@)s X+%.eh=Uf~ejw ;Ӝ+y"% Q+t@p5 q1KWx3v%e GF {#޲PM~~3ob}6q%dRSKnFΩMr!ӸeWYw]#x9±RV>?&LFu :݄D-$vjOިL+ZкZj]֭w9tfw=Yu-Kh&ݿmEݝ'Q=/NxKޏŚJ g &%zgNۦ::r\5(1JNZف qsCsյ,j$Z4Q ~ToD+jxyUz@ 6km0kQua h4^3-&㯃tq~ ?%6r2륰[Jίf7bR%c6(SZ8KA x)lkvd^b4]:as0MlC!fH ^ wE%<X@Uyʩrӹ=:|T E*g꺣b~CupQN- 3!A׫?F7è"}uB{{쩻v[)R)wCKv=UmBXA35GMRqXt߈@e,!]Znu OވAeP@S:{<ϒ F )*;oB!A3cĜE NVFP?P)Ž2IP(yF0ַߋt*1\y~*ٶe䵇Hl3#"ۭkMlȌ]]#šNU><*s8ĒJf!S\*?cU=VYX+w#}_rȣԎ<zKg~RQg}A6zL,c05p TQ͍QmQ'(o$gN'NA:ݧ ;4tf bw ę/B>o^D2އ1Ⱆe–L_Z GH;Ƈw:΍&sߖlnR/kRkΧ{[h6pskX/nWL1VYE 62` FQ9&R05m!/mfR?bWڪ~p}Q_1dԻ8:Z~{ V,Qmm$}|ؔ Yji)`M kf|dKb0j+q񪬪;g4g57wU)/qr8Jr nV,3a2z[L?pX rkl o6u sӊ|ۑ[i0uݬmRT]g5v\nj)2̒,|$Kp3"ا[L.SX,-,xGI4 Gz9ݞdA;i[:*6Ju&ւsu(jز׎)+d_?Unw 33sͲWGf[ڼchyU1uګTI=K v\O'mۇK<4zwd^ms"%qiBMOIugؠQ`a1oc XdҔHqhLvl*-)4>IF_/E_ε:4.+}JP48y{r}|yoE&D(}OюTNܪ 쓘B9SDH!<x5|ΙrG B2c*dH2IK/}0|q}. s(EsyY fw-VZZϪxEOorͿnߞU\m-ֲSrYjALᤀPe:Pr.`9V&w>*%α$QK-9+R 3˙BIQJpR傟|bI6(~ljo<{f_P!>ƠHDf[쐳"৓;nx4V:JjdBʘNB81x O'O9 Ts:kF֡կ{t7ߓ{9t>L׌Ԣa8Fr=K Ǭ@88 JNQ+$xCȸ> vbc0 ({iƇ\Q!X֮š-e.ت[\2̕U~߮oz~P><;f`)n]z4d=ɻ-v.&伓3# У=@O~hhls] |d Ots)%6ińoՁU2xM ( y!+$kL$%F(U@SpV#MI])FXF?s^xa~}s oTQ׏O0 #CÏ #oNzKޟ`s\Jn:/Ql5sj-!X'5GzEeiܳ_ȆabUꗛyh2N|N˼l>6[իtفE}D4dXd"a)%ioT,T|ޞ-b=+Zhc"Q'{!GBDX ٜ5rTvVb_c&ZFY[繳rsyѸ Zn\׶(k߁ao+YwhZR.5g˟5^{:9Ӯ5wCj7d7e_U!^ Cg L_L !)O@~p;qkj{DHnd#UJASJ9OpV(hAd"Kncbm*ټo.tZ*k .GAeZp ^`tUi. Qhk&_C_TK] xxP!)%I:SH<`0'yT"dRaU[Ajx 0ו2h=MjB(`X2CH|.+{M{H|m{n~3ߣZH m.蟽aaeOy渑QLqa'5ծgena|c}ۻ{3䡫a:I ޯDžlZr=~J4@()azzOϕd0jx lWq`Pl69pVx8:ˎ3rJzS>Vu*Щigqo.o>/?矟~_O?L>|OuDCMgO_ϚcϘG ƛ 4ahN\uM]ƕ].e܇#Ϲn?{~i>>p}u.츗BZst:l A̯*k|E5<2 _j р!b8jyh#mMn'lgI)N5Qhf:(DSԚc\> 9FË;f[!sIJVr-)Z, [ơHc] iENcs 16LJcĹȜ&Po}rJ`J^ I`5H4en Bsخ8dL G,Y  EhI LpT]7M/_-l-)~fƒ}*NF}?:w]4Eb򝴀PX)IJ}b_g?Wv}1 b·@9lH~2Zav V+6z=ǚ2u-t\?}o[В]CmB@()_l\ U~Q]i8O å<HxBcp%*:2nt PF0bj+K= 32 {;Y*wr.xCDrY#!Dc[V[\$.iOu(EYB7"QITR̟-ŴoQf#s7KղJ$*\%څɗ, T q!20$MY:a1ޠGw$Y51էsTehX{ُ4kngUYizWhp}~5YFzc9,9Nͦp^V>lA33]35Qf2kRWrzާء$`"CMxb҇UL8qM 1A)Fܤ1xFNb/. UƀvNΕ](,Zbs7YgOQ=E Z3C9Sed6Hf"'ku:{e*BBWS K諾P;YnԣTjװOf?bi.,u]4qRvC\ekAlHb!I S۠[辳o bS&r;4f+1<ឧ}Yխ74> [:6a7{+)qEV{a:x4t uD褠r3g.QŁrd(j\N'9N:|:?ʞ-Wwѡ?xA{ݠr9| V,rA`Wl}io.. Kŭi'1bHHt1P. +zx> Â?aF{ t~\mdJTR^[)LxGhTFLD;, q)&BMt"ҍLQ&5.pEpkiFPdBȝMgHj_>s|4 \̰zVǛ-Ǜ{/_]E"!$sdt3XxP"p@F Ԋh(`#0\AoRJn<[LYh:8- fģ(+BO҉Ip[*Xjģe1jT 6;B1b%g^8U=cvA; 4"=c}q(aFeYZY)t5ɪ-!M^.MI W!+dRfQu%f?3?{IrJyȑg>k]ґGD̲ޑj PEVVvD_D]ʇ;mT'^cګL+d< mA,D^zCImoS߃o]* ]P>oxt7rNqb͘1Ke^D#?* m4 mJ:݆$L{=o?l[mmkmkm,z!&p1D0R@Όyeb+Og;ƺ 5VoW2 {M ) +%I>:&T)tGoR@IJ>Axo Z@>:GAڨ]QBN9CM] Q;ni[j,gljgb9$YqHIҲ:X)r%ɜakR )c9(Ld%1ibYbL (wr`=l:;YW5[Vl" .*`Ab-Rpk<8,$W"*]V[弮Pk0g'U?i_s+vGf2K(viJڄUW^q q(}̶ˮDi}ÛeX ݣ{&FiƥQ $eTafxHQ$ T6\Dv5׮S_S*XBvAyR3bWA*y)Ϙ}rvƁ pcm>LB  մGNj~4[U˚OYrF: Q"KBBl悔 W#|<%۝BF_EvqoI[jU (8R٣-7=l&O&])(N'םnnIrhovݾd& %XOQ~[~8=2tY{ٶo랱=ep5yd1eHчREf)9̉hNz).C5پo/zbk;fN0g7$E,ZUB/"'HZ(nNZq95cYӮUf@m76鷗u{f9 Ѭk{v#uKYɷH W$̀)hO Ijtݰ) 9+o1n\+IQbv=qczFDi?ŷ7-9=@̊fXOe_FQ l&`e*gr̢!5߳ʾ# Xh1PshIh l-)!ub@4Zа{ќcUޑ9Vl,*""(4h1Jo"D5GUMOM{(#ul$Gh࣫Ao?şG޻s#~#MG ׃z_A63CP k \y-$p:<ȒKsj{c t( #~09y#n%3QƵj\r!jԸr!*r!>PlBT:VOh+jZմʅ\ʅ\B(i4ԮQ*֮bX*֮bX*֮btRVX*֮bX*n!bUYX*֮bX*֮bjWUvU]kWŪmbX*֮bX* .eL 9g8]):fhے-h QQ"oL)|LA{D@ց,>I:,9 HS(8OyfCPDrȅ GZg- <)5Y8!6cI Ύb+W=]> MFh̢/т EDJu|D*Ţ 2z8,8$0f V{H @&XAc8Ve9$eIBL*t [BG0)("R)r ]+f*>e$W\/Iea޵;GY*RdHFء$pTzB5\&)1H R)dxR%-@ε~dlg販gjls V%!o3D0ԒZ%oԺ^!Kߨ=IP+CNjtx:l°HY6+6Ys& SgMYYs1G2r}<-x{n]n@;Nwv/[93;|02;";8y<*1@ko%г/o5S eۻd107T movĝC`!/@.oD-q<6* O^z2*m-VHUTSJAo "Cc0G SEde蒭VVi=?D㞏Έ#Pރc)'gʛhsV%I:U)3ጩPbz-Y +migUa,MdQXedM],4=)Xe@y{>Y25W*5hJt'H;Iy >~)7e^Ϳ~ay;ɤ_MOiizO yQKԝL{Z09P θdX#%OxFGmG8r_0i*]2 K'nm#kPkZ >O>)Lsi: S1~krKm$UxNZI%XfJ]_M1% ^m"[qZN>@Nryѷ]+?u~.~680+Ԟytvr:_[nVDmU'Ox#I1$gz'=F>F )bEG U,b<^.gg-/'}`۟|ɾ+aCS%K)$>2Քܔ@1?<ľuj jH&1l<ijS?_>?X?}_>s?ϿY? 0a8ꛂ0Ϟߟ5)1{??vG󖏖vxCbwy{{_}Ρ=+)~a|6ޗ Orrc5+>+lԤдI;o96q1D7ܛu7X HAs.ǤJJz Arfl!j-+CG!Q!gϽd^(tܚd@E UYtLg!"DpFsAu`2NJ=t}$ܼtlUYY#<̚*w[A\J'9r3~y\VEڠkХSfl =a̜-ң!z)v@Jo!ãQV0cڒf&ɤwe(Qt:gU.*FVt>qpi7-jsJWWԹRx|;q3M!5.{zw-qG?l126e^{\|FKku)*;ۤ2[czo gz{ H[n.*=kͺwkw _f$ƍqz_˥[6R^M!]a9'* s+m+JS |'JD3z*>[SfiLL~z V_)1JV0K!'Z* pFx%ЃD`97Kry{Ao"ٻ+k{ȦgCA?&ARV_=2 BG. 2%}Ѥࢇ4oI޸8g&zmz= Jzg?hL9budp8ɭl 6 6zR\rQgx/]̛>M-{/8L9d,Z%KA)$ jvJ[h:]e,nsҷ=ыoš 0d6m}q죗9 Dߜ9J ޕ6c"Sh/4!=40h %@pu ݒjWëŒ2eK1R/YC dJq!D ; /%E.AWiG "bŃf{mQNc0n]CO-@q׿*cP}iQ^F=Z|tA:D@.' kZ%T-D|DRH,ZJ9,0g`+ #y!Y׃.uv.U1Hۜ%:fn szZ)KU7_!*XހPJLk/Ug0SiZM (YQGSpd- v솎I0L˵BƁ&eçFl)Q h팲qpYB~šh)ydΐ4N &RuB܉=΍.\79st2_ݎ|:݌ψ'mz 9v172mD/w;"S%Ύ1g> mv*z] b#%{f *jAJ.m4fQiT+*.]tjŔkn{A Kh ~xv*M[|X,$T bA02ԐIJǫ,\aA~fpp XWq@7VNşX4;a4zÐBo\͍\fP{^ovº<A?l5TowާacKɺ*?Ò0K ,Y)I M,Փzr\2Fpej]8<- WɜKcQAsz>10NYnKʪZbw7U츩ޓJ2Ӡ3cD3wBVK {\ms; :Y)$sz-4k_aӞΠM:Lyg=քTL뢺ǁAswWM*结`n%)Sȅr=rg^A=j'jhZF-AS"K$H# r2  *{ܝy=J1%:5זp rrfDh"*ghIb`<3e*kFLseRKkL^/Dax6ZkSEfӅ.T.$J!R)NsҢlEao &nXi~V˻#T]{շ :ŏ,Z ηŔv7\0i f8)-c <7wRZj)ed2`sH\)Ϯspwcbίnh8 g޿[ow#=4уjSW? h1 *AnGQ?O%RFK˅#w{@:TyaI!I}B_/Ohܻցz eހՖ. $P)RݧRh+T1nEkbNL) s ]LBUat?OU?NkLoLFiO~ @ >><>'v)'RYM"-"-"-^r xXŐYmYmɎնi;w"-""-"-"-^p ϔI*bȇj8 #-s"-ڢޠ -"y&%"-*&"-"-"-"-"-"-"-"-"-8 -"-f#-~r,CV[GV[dEV[dEV[dEV[d=N Ymk"-"-"-ھkģ(cYm -"#%"-2H"-""d}F}!i4p<|G~L1s$hjzt^oěۇ/K wb5xEzNns֪pKm圂d"CpL8DəDI̗Nmͪ>}FJ&a)Ĩ2镠631-V1SA?izX8u`U~W 7 %L1a͆_E#.wT c[s:«bݴT9[v1E2iX-ͦdc 8$.zJ碽%t~zR6?xh+T1Zpcje+":xM-TBVyp68kI|NMNiSҗz#-c췌Qʛfq-mE[xrB0euz[Oag\G'ڀI'MXj(FXC38!pED-pLHZW8ht\h4ELrځZhXQ  `B(eJBP(!C**T6 7ؓK1Աwc2J H^"Yg͌6F\ ([ؖ 44ԴSjEk bOiջD퍪.b;Ἡ')Vٜbf89+ O31_dA !bD HHK \%+I\Wi눘3#b΢H.E~n.YM#%VYʒb\'% *ȕrvɸx8f;z">mY]4g۽ žޞfSR c/S)@UwN!{JJhn';?yFݸ{P;PT~ AR]urw܂ВQauJS@rD-#ȣN v{xJg]QV*:ȱl[ɔ9N,&9w":4Ci,sLQ#lT8ZCj#%1i_m"mp:GyS2HceXYa 6lš--5|:>D7iZKojl|DWr˖E2sC措S\ttH`W|䣸-Ghf\*QBI%e^b ݚ]m(S$XX hCxY2 3E酪"OQc"Y]]EC6ΊIVZӂ&d%5˥~(,W03i 5"Xe ܒr*ZHB xL'XT [;.el-fNSXufZ% E0D wD J96a9s7Z1|hSy>׫v*>떑f.}O3x=r7kBTrGyx ÀT.vZw_[it{:]ٟgs6%RwkzxA0s@K-wfGW??z:qpI4zBB4?ozҖZ(HVۼf1T gvq qI I`!ꟍVF;,i,%C^ 7T+Mr i+ *FJL[RΙ1nFu1(.kK꤯ejֲ pE5ŝs8hws}屬\n +ib8b_cU5 ߧDž^;巟O?߿~D}|_0M&@x4oyвaT fUd\[},>\-̸Hg!V^]I]|g_iTr?Jn|B`@[ <\c(WRm&'\A7'I*$ )4($::RJZ0 HHG%H o8TV4Eaǚ1W`L4Pdu^x*]U4QrJ u:99jWRZmӝ#a)VȚ9%GhBp^^L?.TVⲃˀY,C2n&V3a6=A9k 6ҳ!Zk܈'v!.=LLrFK0Hx3jS}2#,*5‚$mHc `*`<,jv9w}_6 T| bdƅB8+o't|w}p=~@4&㞗Bw|i[VR 2npl&XNiNB ݜܔUZ׊u>#$PJ6nvr61aW]$)٣1JC,LvIT7C=%3/4 P ⌶g+T\~'OJ`T),DdMƉhA #(H8Hw@-%zr`h+k{Ȧ/ wZU_1ASt%a4UTsc%bD){}P8r\.ä.*iPvRGTRJmɌ|,!&zG}Uˌ4֕bƊcf NE9~tt v =65!l&4sϯblS|V)8gůa4 t´<^"M1KM3u>MZV`J9$v6-Vz)e+J],(cP/YFN? JSvQ?W.gVmi1}h0 PbZnZ~ S] zyV&3Py>Rk}N#qzzW8.y+;Vdei*-h %ȗZrV:P(M;db%%/o&;"y8")O8Vy8-޼{W_iqW"c+aU[[s 0w]Q`?L?(oL6~'GC.cS2.da) *`1I .pTƁ+WlIAs#YD'cBNTt͈߶o  UL߫iO^]lF^.9>z@^3Lہ~S %t_Z?G\!x ! AYS'RX Dc T鹣Ig1s I,Ew,|я9ӯ*OUM^ڈO$݁\vGH?/SF_|Px%BLF4xpJ~jG@W`&hPBϼQ]A Uh-.p]' !9D#sF!9Dk? DC42!9D#cC4rƷHrTrTU.GQrTU.G mR3%U.GQrTU.GQrTU.GQrTx3ޭ#;d]t7x;XH>ۺOn>}}=&SĎ[LiF<͈߽ %7~Zޏa;@:!*THCKK)FD>#I !p(u{hE*L<)Ɩ)18!.)XYh6d8#1pv:8#0RfK̒fmb7\/6WyaRùBV:e9o&Odt<%7dɓlpP?r*:@v^4}0*Y:9-J-AƒEiC5׸'5vS)Pʍ"6T<(( -J))$1#]*ekZc7C>8yFK/T>dLׂ4sL^1k le jJ]>`h5Z8gBrIN2H !p؇1=ԊaK~:1"k^ 4ڶSXcodlٌ-fuRz:znca-ˈNA|ieeH;eqP<Ό;(H_EFߩ%'HBI|~;|LUO Fmz وΗnr5cb~}H;tsy|q/NcF1ZleF͝QFGR-:dFe%j3 =<;YNr lv`MJ}.G|$NhP\q?Ljʰ$f>L_&WXoq1ЇANӁDh;Zo])Ogׇ OZf{׽\,O~76=ȗk9FYeYn=+^y޳: xζўbŷv!pDth`eN!Z;IyY !051DK}SGCJ9MiU"bK(F[ :aFhNPƎ1fg$X;yp9z9`@ƶdVZXNO'hӋH^=9+3^h+D_Sr<ؙ '$H+7L$t (W*5dYKe.81e;~ѕGVUF¤]m9vOxų i$ wga)A@C#R !iJJpp^tU_BJAY&D4rfaIt :Ӏ/УA{smBؗ^ D{2;=LWSutc[.CTxgGl(yG#µ^ NT4iG-QQq鬍>Ũ=V&|,I`҉g9;8j9 nJdJ!Č*-6g$DhVLZ"$"MTR#*#+TT!6Pd4X:ΆrV՗@fgk5E @6hLPPKa NxR+3! <1`JЪv"-_+ ?7 -0[j6Sv.Dyf6 t3b0jZBZHܵ?P"4K×ch#whݚ6Ը$mϧK|3J3[,;szuT6F)Coijb (w[ %QMܸ5ɜbV?ijG=/D!|t KJS(Q93Z e'LPA #F:*Q)1u?8$m@x 5"ZT(@>@UqG}LGO##Q0#:cC H3nj]V Ţ0MOoG.M/4H~񸄤hO&!@T #89 QHUU"=#I2Hƻ BFx@?HNuɰ\da&Ğ fV?" =v4?q7abQHE589'-Iiq_FY?c??0԰ҌƳ&nx0<r8 zQeM{0pF5:8Q0J('욜](9NޞMgh?šiזkoP7{%٘ qpcn O]rt)KQ)W 14ůM!W񶹶!Y쟗-,?Uw]|rȵ"2ٜZ-?|4~F#t]=$~ke |R{Ef޼bk݅of>-QLIgm0J|n۵ 藛lG.h583w'#{b>݆ކ3{ua1ѸQ8yJ> o=`8Swed]w׾gElK7(>O|f+59P{Nh37XשfGe_/Q?~ӇS?ǟ~D/>g`=. g}C`1<֪ԭ~[s {Xɧ^L>%G;Iځ_qۏ Nrj8㚎z s-7i$FR΀ϡ3!1@ɟ&g'Fq>}4$? ;W@%4%lb:PK1|v(Z%Br"g1bjKw!#'~Ӕ-MZN,4ݚ 5xsv7D$mk#9*b2F=gwG7zE4zV̚5%Y3]lO.ʱ'8ĩЪZDղ"P.p؜B3.rl"wšŗqQͭĵkDS∦I"4+iLbr96+kZ72AYKTPԥ(~zIJ˅ D V=K.iB#=XEӉ%_37,G^w]n;6ع#+==a_j]uϭyttjxUޝuع5zk^et2W{-$Zxh.t߀ wm^ZFoeN#׆OOTjRa*5L0u'Rԣ*RT0RTjJ Sa*5L0RTjJ Sa*5L0TjJ SajfZfPTjcSa*5L0u#RTj,RTjJ Sa*5L05n7cҕRTjJ Sa*5L0RTjJ Sa*5LyG>O./l@zArިTOʊZcnh]ԶuNYӷzWt-dwlSd=s|cYING-ךROmZ 1Ax=KBdid/~/M^6_oYt3 b3|%3lvڄϚq {whudcveSО__;ไÛsmfO$|^\@3{VTطqƴ<Ç D88kJ.sv$r] H!cS @Ug8 ŃiG垫bŊƊiȑr{-Z>@WY'eCtwa|0C !&PvhøA^o.V <^+yQA KH!ipML09uEo =Ce:V()5{3BDN Hzi!Kg˽̎dS19F;AZR0I3cSF nA28%x%9qq>r :| mhP R㯝vSx91QHC}jU!UkT) R_# ؓatV;ʌ&j/.x=0iax2 L?Kiu\O^ .[Yť lO׳eA.Mr7>_6= ;Bf͖UzszC$.}b" jR  ڽb% zV$2k{UM+oqܯ,v׳U [<~D;jyF_j7?%[_/;]5sfVŬzYW ѫWV/;Yq߱Nt|a}WN^JԓҰygAZmKظӋF jm{ få$)]4v=tM~3E)<#\[&;.z2Pk၇[L^z{4:6a?]Fs1Yqho$[vd) <褢碯-48߽1ʛAQӺp8-8)}zy5iPrsVH/=W"1DMh !h$$R&hMW~ >7LeƳpr4u" l:z[K3wd75x&|mţ9],_K[~pSG#rǀhdcc%CP&qs-K GԄZhƊkSЙ+^+ mpeŘF&J".v)٨Wƺg2'E>BUu`9rIL# Vt=W.}; ڪc-ҧFvݻ#qQb' `VuA'oAg&)BM[φ%!USLV..J0݈5z+]*OX[o.dl<˻_5={_l&8;^;BxnDdv&L-QNh];#"ǡ5jSi.6P>m8|Z2)ՓG] Z\I+1*EkCXNAi >rΧ]ЇYRn˰ĞC0oܘ ;D2=VSU1"^aӯП63w0`Uz;RTH@܈Gtx(#G o4b᱋e, rWpk9]0Ɣxu=Pb^DÃ^vJ U&!X,D*H1T.+%CDhKFT–ntfEgQlYwPO? `6^$RKf/ܯo =2˛VwS/}m*]HpC,U]'jAq{ ϹoMrc bX]$WbrA(\s%BV!5,lRQz,)ff e5rTdvJ(6YmQqq&/Sk|PZhdQ &*AY@QR$T{.V`vW`fŷ(j `ۜ*c1Q\8'Yc5L I] <-v{gU@Pޱ6la "&+LG^`:Ziy6'mg4* *1EQʅ!;Ϭ=Bh!O0[_ ٷ~{A>a"9 єL4ՑlZSE5:D6܎b^V~s`*r_ &z+,jF*dBZ9jNVs:Rc% >PPX0pc]<LCiN k}cXS<{䐮?d;4`tgH$Zz&!a9ě`ݓ\KTs:*]XcoK5"EMZt] R_uk Ik]TБW A< A1e?r8- aܵ{];f1ῗ:6]jdb*$V*璘R /Зչ1QJ'G m)PȉnDL~y7viw~vt˳ؐ/T:buP-w䯢r4v?k,=Tu;,M{[<,/zm_߯b|a=p^V [&WRF3dm94@ѦVmkֱ R18-l[tí%ؤ_ߺ9OMCTuff4Dg吼 B 6Lhrv \QP@a0*x^TZmtM8IZݠ2#*Js7f WA=&LO7!|9x\EombE'/Sɛ'vm1%^ D'3;W[C@(KJXBNg*= =a͹2+rQjSRӍ69LBĒ6j v!{sg>'*EQ4e99,k1*F6.Y`#-d+(linzLCF2V{B56 JU i.C-nͳiD.a`F="!\Uucbam U@!7zW %/N܇9WU[0ޒml6H+VQی`j3$Μ/^-&-QC2`fFK{LK;j {hg{Ph벼]^簛nW=6Ysn%" jiwbh3LLAhl"L̥vw`8[Q.:e3=NZak}; h @Rumd bT!HM:r'GJgkShLhq.5RqHoO}ΙׂՍ9=HGq:VBR.S`4v%d+?hYcF֧KUYgB2A @kR;낒55iz-n#ޢEYu\”tcɫQ>}k.ti}kO^,=܏Fj J^[jy<=k*?&8c꿷o.D[R0ijWd\&MN^Љ8v&.]w&Wwql2'ͻ|fy1.VJ`<"ȱ׻}l^w\L>.ۥ :|/z;OȘZ$vc-lb}dZd+dB,>N3ige4{'i7h{NOqEPk5w&MezW[ZxxvNo&}>_N%w8iשGoS$q){Ԯ.huثɌO 9yyr'NArIV}*ʛ *YlթyעזT||t"POF|9iseS&Z75(wJL~&^N; "ma]^*rX߿rN,_MdnW횱 ekkS;u m?)Co*躼y:qz|_5)݈e; #&ι7z:ɫa.D"r{mq[oϋ}Ny_y燸0H.4*%l-Ga-+3JS-Gʂ9eJ m]a&jY.`;4"Az`l;LʦU)*HJa6V!˵~ˍ{1/BAsck8}|,5vHE' Eh-4UlT.R0^{l :D6)DXrFHhW(Bp/BEKNYW OGh Bkcجg .fOm4i᥮G+B[K)^ FEcrN-)iV{hb $!0!X"kE-H<T<(B% $%"Rdy[2Vo4ۂH7нݧE>듖xM\M NJXM0dۺYFr?9گCD`.P_P_u9WZ;.YZY9b&Lr.)mP|x6.7b)tү%]@-L(A2yZ0wӁU KDFxЎը1Z&Ͼ jwW[#/ҥ2}>Zܙ"s<DžCo9FKuR ӣ_8e+9UBX 38SyWܕ6m񦚶gzɍ_K6@L6gd_%`YqغXՒ,Svkb5f,V-毗qܬ8be./zbsl0xS~w}ϳ?Wqg0NѱBQ={Q-h)PtOC1jeͫ4y2g?M~`:1 R+M Z^9l%)ahD+]ɐ~|2!J$G .)$@g!әB^c+xIF=b3m1O]78;d9n>wUmpj~:oÞRG-5k~gGE*.f1T=[CwѽÛyg27Vo 5.-ӫM.zC/;'nOrG&m4ܺ#PT0^cP"GȊi+U:V;鉃>xrr^c'9 Aznl $ rI@e$gSG'?LIExkߺo?:o[g?۝(eswJUD_ى=cwO*#m=8.frz2UJ>-Uh JBŤ){AHӡl%JP)°I ~TH欗>X#L^heSpv':@ ]=]Jq4|ݷC0NOH Pg{G73@PHW<(pUe7iUђMwtȧN|nge+ϳlz OwE =2~'ӫExCErY#TcFxI.Rg:EKbLx--yRPFe#jτw77 >^;YI7ә7>Fnvj*u9߻|;h.{oayQq֏EWkٳIЂOa0v Zѱ#Q@ܻ}tn{p,dn,>W ش{>gz$n$dd c2x5{o50[6@ [lf~PeDx*leÿ{dn@f5ϺGh X=( C=&S"Zf']zI-Mna -2獞w_i庤֖{M L[8BRVIBab<--#zy;{fzGOI5B-hCʙp\d[eɶax=树z&&nqdha7lx씥J+:;סrYT+ぷIDzO,[#,%eR.sQП{>Vgݹ8С8Љ8+J^R񎲨()30ԥE9I@Z؜5@7hdqq,<҂F00,6}ܥ\8p!^.<]?Xݩy BmͳH}bRoO cF:I)0΋Id d\i gEY+m5:ϼ qǮ'4DDz(TWՁcidJ&S-ACVBRq48K#A9n(Q\#F\E QNyisA䤎GvØb pȰdRo bM [jVо SlM:ڔ\2Wm)6_son>_b ;FqbpfSPR0%"i0!ys)M ݸ4'(̣#rL|7"qvh2I.XHadP8劸(* t)p-7(s2]yqF1Τ!9R 1K Q)C%cDS-]^gņ'\l'EPs`i(ϓɃONEfQE2*(N d" rJd]x}/?W{waGBaEɗFRp:cE$qϑQ I"RBV1姣O%x#6@Hd""."!˜AHTؚ[ ђ􊣌TV18(TK~]2Tk߫(nQ(9I(7N[qAN^tR&+4TH4mVDplwX\ b"r(La-$aYʤWfIyc-`Ԧl0у0h :MrX>t$@tR>"] aiP!ǂAV R6) gfR'Z)8TºP—wy%yvYQ?~:'֠h)RXco %8Ij dȈ3Fa*oQQ6k!60ceɥ *GI |L+]XcnnaŹ -%.-z^󊤍BˣB5"A.2%r:Tu!@DdH%ZQP/K BθEvEa._?}o4$A@`< ҄4sS1FiG2'!*!8Q=頴,H@(72ăJ`H"9!$rAL=Ì'i-ϿZ4ؙ3Ёl)ep7k`y%"N78yu#N"jҌ*7`og?x԰2L @M ЋWU&`u^3QB9agBI)g=STgqdſZ䍣^It6惡Byt<\,pBkx9(y/ 1 `Y_Ňm|pfm_>}z;őkEd޳9{[.hG|Ime |}R2xLփY7;_e̮\nz9*90ԣ^ i4-7s@6.+K;^N( ;|۰pF[o*, C4: g0b7b{-Ǚj+#xȶ=+rCw |\`$Xfsw54 nfKkzTPS2Ygo/QcoJ>g`=!0!U=pn~U[5O=}+K>r}GGIoڡ_@|/ yy/jΪlMGu= a~x˃a=X jFgRQШ0"vnǼI)VjMOnr$rh:gzʑ_ev=/E`z̼L`^R +Rڒ߷/d٦l9:$OUbPf鱨B%el(d 뒊Q ^<4y x,%X@ P ZW35"$5/ya z N}*L&Svw콇^ ̚5(~2k&Mw.4ʱ! cաjrH I&>rl{8>'ȱ!MYP<`y-Nh)h)@Zd끔+˱YY~#pTe'c!kbCwyk6(B* *9:+&"]H"٨SBR.١^F,3b2em ͆XwSY2 o;l(D(^O09IZCBYKu%KƖa,PKİWPATrٲF:3#X$kE!QZS*cb,9leZcv8+ ȡHgMff{&N߿LR%Dsc7we63A)hT/e3$)V,;5P%"\B/W0, LgfB_i:DY Lu6(:0 LdSZ\(Q’/[of߰Xhztj7:8]E>K=PIzKSg ݳww79<ꞇ9%BZwov~ymحp9oioyu{w΁>5. ҽYP~sKA|De񵎸=rΖ Մ7Sg4ӵ Jb L C--Z%GzWVBǓ:bl\W3٤tBQhli!i#M-X Fg1J$-)\=4I@'!,:?} RA. ckVy ofon%ܛ`煫g>.y.J{.(^]/J3|#v<8e=xQB8x9 DP);gZ$jF(ic[0t i(9)g9OwKvYzty )qF ,sC<}oElKmvT_h)jg!H8Վ_h4}uA&L,_W^ }M~n16Ա^۱DFSysڣpx/f| `B{}@h=Z7 TCZphӡ%i @)# $|M]Qrl{P@ѷux# N59ѤeHg:]4P*τ=:]r@o% !k Ԉۈ}nfݹ~(k1m+|6 1=+u9u8b~ g-]*$}%[ȢHEI"Upzȵ/>/4R=9^uW֎2 @x2 sV!B:>?O[Av clP {eE2I*5joBtt(l>-x%9rq>p .nf m0{F)۾\unTTJ!&I;t@e2&vFE͗VTJ)*|E)ֻ(2#Y9냨QJfiI0HʡS۪.d9I)`,-#aBx"Rh BJ)*j8ÞjCЇlRח(ʿhMg8"8[*7PEQ|J(Y<+&)]2>d\䒩Z(,*6z[ze"|JH#XHD:£H%]2X:ZZH2U SB{S" C~e7A$I|TٚCdDVYE "N P,.g==jtIt$5S" 8 [Dt5$5u@[`G+zp}J3x#)!m289+A#MNITEeGS%dXB쵭]DVht/yLD|-˘b3 bg Ư_dLZ12*!{y֣/@Qh߫G;XUCЎfD0vzԅ5j|K7eUR P̀DY&$ sA{()3+o7Ee&9J!+ڸ.,:Qy@.r dwm릋#(ߖOyBiM{uIk!0΁Z]*e؂{FJSho5 U䞼|_ p)_O^;#>OK{^w* 4\>sd3@ 64"ɐTR<^Qgd+0||qn:MbovE[AA4q2tю '.JC*#T;hBMX:(ڦQYd222YH ('Jj͆PjRs (sՎ[oD{FN؍ezrԪ̃==ɻ>̯o]GkTLQd#NY)d0ܕFGp uxr2z@yHlG {Z*"#Fvȃ_Km-.6X@<&%:ONi!Ix=b'˒кW雫p_ǧ6;-+^tW_РTB 6IwWC TV wsd?X"Q^H9"L9K5X\LtXjS5}i;=Vgł{J:"6Kd:<(e}2s?;AFY"u>4LF6:i ;J6w ^Uu饖] r6PB^dau":id ,3$cpI +R[EzRɜp}&Q$0,e?{VX-1W*J>Rc {oSC kUNZK0C ʩ_WS_xjt.-?"4C7Ue|&Ϳmaye0O/φoƮ뀽t(ϲ&n{R0p"(#R uY:$L&)Ll?wk|^@qI8uJj+Oy8 ^N?V)Ѵ_Jjv: FePmȻ :fd1K<^hof޿7:-Y5:aI.rOh9I~f5_~R33-o\!: S}~مg>gk㔻OQV aQj:^pF Ħ=}ԍ؍Vrc7gVLF Y+x4MΧ[N+W%|MnzVƆ^'5OK_ՄW@;y"y5jҨeǼ >cq/Ͽ~T}O?G)%^N6 Ag7]=kתkoٵ\9<%reAlN8Y <`YJ%8=d%K%lea챚bUů$=|f/-}>?IzЎjm/Mp?g!=^ӪQ]\O\_Îj~W{hV-K9գgɀ;W1q@LoibVGD/hs?IAQ|k(tB,R2Sy`uG;䵲NdU|֝YwmnYO_:?o-ΉgmAMiφ@l݅l<S./{/a'qDPWw H% @¸RARŔ\^\`{nGYzqy uL9ڜ<%{O/(яصտQ~j]7tۍT.-hrXUN=Y}Wx7NMZ vH>֜87'?ʱӾiVY=`hmCvTHҶ/APC>j}d`)S O;@-T0@,qla#,Mux(#]>:1LDffs y=4:4u>tѐx&D ( =7CFk aRm}c+mf6}5ӶX ۍk غe٢T"[pPj%0@5RDS W_}/4ϔ>UTmѭ]d@loL_j c5o\51Tcs{_$gShVXt"t0_7:^/12<@k fˊ A9*|J%VH9!f"m3pqh4E,VÑ-?"71ZBjO((mnPǖ򬋆ՙ(ʼn;։Z7>tʱ줥tߠ^#ˊ42j%n="^h$+7$+/RR-KI^Zsbl|3wMzE)yŽfz~ҥqUO`MD&\Pmp(T*#w•C}%ȡ%3:>#ǒoe#DVjJY {4ZST SCRZɴ4#*1|*92g,\S;NBfm8$ vy0u{Wb@6{Xbٌ_v]A[TLuHOfq pJbM8gb 'pCFהa!: Yc@JJ0 X30լ`ͷQCo!K~jV tp^ޖZۀ o ܲijPHaD:}!9)xd@:>0JIr 5>8^1Wι^JwM/xK3'ϲ(pfs^ p ǜr5[vem4ugS㻓OV2 ْf:+t&dn)kDT79]ػI 6ܐ OrRQ5|4N^#i㾚3W!jk2d!Dt׆Zsм8у|c5k>VGIτDj^~Lljv;zx5QfhhV7im V.)YlG^łZ0\լv˫Bnw_°|7~jaNu!*"y}: s$~DY 59Lfq OADMzˊ-Jh|*HTyK>d+bBK<㈲d,FR 0ğ[h* ,os<% c´%J{c&Qt&#ZdRڀ\@JPAmA+ѡâC9Tڟ.ug7|(aݘsavL&ahCǏ%y GB)М:G' { hpGnԀܨA!C XDoTasT $h˭{,Gnzg#7B =78 @SGHћS!: ׬q^ubtHf| rM\ ITdRl ! *0JVaA@QPcɏ )(Urp 4c+ Bay89xσ Wrz6LAe4Km3ɇl_7 ! ] V9F R"PmQT'2rXCPRmQ0[X0hŽB('TKR`<2AZ,HDG m85x'zz<WqȦwW+B,j$\hG!$4z|Y|P_u= gQZ;})D3% XO)``!LxɪڪCw:A g"@Q`3cNP4Z4b"&,6{k{Ls?&IQڝޚ j h(0||;}ˇOM/wl_F%Kǯ?F?5h6H-`zŷRJ!K JAΕ%t([cZD4.ksˣb3G-!7x2Pc6@aMt_y8T.@x&AKjJ*/8;)wב!} gJ|ٻ8 Dmf_ۨSVߺLUꇕybWʉmx ;lj@@1E_42"Tܼu`}^)5ޏ>c0Nj jP5o\51TcO7eK!~_g+Mʺ{s)>N]$,u{#^CbRrRn)X8DYeEo RoN~zhWQV13Yaf$GkH[twiQ NSef̢E1ϧ[u:b8R<N<ͅE]r,;i)7-a`l̴qIFϻ,ŝ+7$+/nަ/}בQ~5[jU/ڀqFYfXQ{l6sMOά4aC0 pi5wlflzYlSަǔ42[]qth77xJkuDtRױk"{9$۴P. EʎCʛ)p!6Ld4s(Z1 :JfrK X& NB541-VB9Ouffx<l޼}xezU $QA7YC]h.PamysqwnR⟧^LݙLgG pIIɐwqF4O-fQxmgml @CQ(sOYwbϡsZn>4!⯊**x]ȡV9J;eI׉G{cu^\t!;JfJf?t:.kyݎI4YF܉ҕ! (Q!.X *;lmf[m:jpA8SJu""7uӆHO.‘RqH 9^;Kh5# c ѩ{U ܵ:趴iҵFQt꠽CVJc2XJHAM˴$re;X!k `+_Ŷx{]П.Y*DDBmi}dv+Io0GٸkЗe:HUB40z]2 )wxoZ*}S$H>ۥ1&ȥֵ2`JXk٭׮Q3=?=of#,OA:Z2H.% K'k"hZ'Y'3" R $7VH-7A`-ز9[ h Q9v˽:kOvm[mwtB &ˋao*^֧"4iVÍٜk.rsߪVA~H8!)*"0($c. NѪf4`sJ9&)K.H`ڦH䬼ds|.iY3Ff͸ &CR)F.Ke` LYвn; q7-xY: :<(3 $|6Y;BYd QsQd [ՇYc PLȐL%] !H:>!eX$o;5r6ƨ_,,q_4bkF\@oI#^Rӈ?FڳdU bxt`S0BsoeL^ K&$/djD1cƨ.%Q'r3! n%BdI(Jo"DղFlya}u%E޲^^^ݕdz>|PaVѺ[ݷէMJ75MQ^]w ZUVy&z3~U.YcHsb0N!Ib*:D8xJw N=j(rjGZv Zg`w k=s t BGZ%c@IsC ΋(v-G7[Bw89C+ڬsxؐy ) +%I>:&T)tSID%鏕JK}꛴Sdq|@re흮%ۥz3-5-Iq )„[n Sm#j69 ˌfb[<Ǩ=2Tbdt6X   "Cc0G @ųNbzd͐fBYh,Iei "CRI(xN[#gKoM2)pkѣҦm{` 4$LymdD;E@ZDΘVL_u`iڶgad#H0 k̑`A܌QсܒD'A@#ufm/Vyl"R d+a}۰?;ZDu3 j5 Y`d Iߡ;c'a v[ښ$VeP.ѱ=C %(ӒQh5m avԮ (9ìu.SAy?9 3CJىR,YH9X k'@ɚClū0 \MV<'8_iwGHcfgQ{WzqհxR *wSGdsEVT/WC0Uת3.|HD+jӼFN/TKN(2SlVhEw&]SῧWj8GՍ9[Wߧ7o'o{^q7lmZkb/Ӣ'oN`}Rə#)}H竆+:|"AL86:VbɇbWc{5®U.&jUs%lu\<𸐑2&pK1\G!?p梿ǚN5wk:5uWwqJ7o߿+z}{ݧc.95:`` <D,؎ C۶CKP[ D/66bܻ7m/.F\2 c9\DhW~\]^]E]R_!IkꍫEܡʺ+SB0n@`UrH}UmtM|WN*E) d},2嘔!=ECE0aŶ)bK}e#"r-ia~fw:d^ebv_WhIhW.J͡wzA; "#w%%ObR1xQS3& [sҕX|4{%/ Yץa|Ew ԲV(Z0G~!v# /_E& is/_EF"XI.yk)?;U;>ZKBjF%+g0N&smlD7d6G3Ӌs }cx$gL5Z7=O ="/@~vT^A@f{Y}&߂~ eYUvSnR2IA/WcWǀ4ptft(M9NԖMs:_!)2ҹT Hݬ0V6XD0*(R=3CIIJjl@u:}nkH;&GIoS 8y5ц>pplox^õ!}KbOx߻95 yz6g!/msۄqKUɰ|B>B*H$;+vnO3p>/jLon?9׻3WsY*CIĺ]Mgtq]H,aS m'zlj;0Fa)?7iLmT],W03=TH-a7ycBZ}uRlIs4t 5hrʤC5\SV:]Æ}ʹRhUST>kX$kuiZKm\&9j-\;qK|I5K3u/j:7f()n"7M)cĘDhY{t 13.c7iD34f"|cs1 EG B%ލϾ{"Z#)te_:qʺzlqih+K"LEt(ιR!(=&xB. mK>w8gEZG(D"_`=^WH#ULGv0vU!m [?GQ+zsbrX\9B Ԝϛ&1'PJj3k9PJJW[u!s2J Llt s5wx7EעֈSB\RH((E?`ۄ ҋJSFl|9cb)h-a]G68Y @R*C!VeW+q$spyt/Jmlt%Of9z=Ҽ1P k96BePB kGU((JUM0RO)c_$v y .U!\ "U2VjIQedb@fC6l\7I9OMtPIw4 V2TS z,v6PL&6@@;-V(!Ȯhy^.J͐jPoB+~0!7(c)hdPi0HY0("29{$INq1:B&ox$:jUHd+Z Nk̀z=biDF9zGzk9 8Xk+$[1qE<pi8m0޷4NMXi=]0 J.YyY;or):btBe. p\- &J#vBѸb!hrt)e`j?oz߮n-n>>>GgsI/-eǴN\S a W~_ytf |\Nmkl8|_|\Z~NN L=P'LAXZ@ V@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@u*㴊 ?PgK김:=uPgX:EPX uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uy4o|Vy>oҳyΰ'@AfPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPɂ: agz>̵l@XOԁ^@$!PG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uPG@uP:|Gvwo^+A9Wk߷?ްޗfYѰ/ 4^}XϿ?R$#Wtn\Oۣn1o?؇hW11> __}Yx-ߤ-Q쨎/]똇NWtO55TC1e$BujυG燸/7?9ڮ.bZx,/8zm0rZAaѰi-ݺ\AnsWU{Ro3q]v_4('s鴟.ۭMoO(J'uYᏦ⠰Ư(H2cT:" TpSq69pᗌx?)uyre}!BݏM]:yyxˉrmEr;kUCR .`4j*%W| z"M7Z)ոOsϘ{w'aܧu@!pײ| ]+ ;`ZIPj@٣ ߌMHqH-(KJӿ4=d QL~P,(!EUz}ܟvl!5cLVyHVo'GU@L Z*%WCMU)U^C=I@ـ/&?`;|)~˙yY6|<)3&t/sng{XcLϩ261pO5\lYS㯮#mYEE_mU/4`o`ih(VF+myVyԶH(~xqz$xvp5/fZNmf_vs1tin,OᦟFzBLk;^[amZ'ٯ.:<;zG vPY/O_} ǧ6|C6ݵYj,i\(_'S~J|8vCޑA~rTs`R 4dן>~O?Pÿ}/?ar0y>y5vA\Cx ,~QVf8zjU}*e߇Jh4UmMa#M㠍&N$NRLkCӁh:W21ɇKâ5^$RMpxuR䡊{uVMf׺$Țp6mp;Vù:9j!YɵЎb3x&Y^ Q{m$u[U|GĖ4 iRK+|'+,8pơZ,NqP :U"TPy_c^XuV8#ь0%OYddͬ>ԟ4](cOq()A\R "'Lxhcs 16=]9s-?DMdNX>9%hfeNh2D#aM5 `$1YI ;Z".E)\-p$)-RP6z\4`Y(kȖ/ Ŗs*d@4 [ξ<%*|YC:yO cԓh4gsgXz0R4!g3Pf_YzMZ:dSzrM2rϖ!7"#=GBS˚\'dֈs,"3FhoYﱬۚd#,W*^|$77 _K3yCI''c5K&*h>[qqW uɆ,rv\-Ѕ!.Llq>^MT 5gdqQ؃J <v(r_iX÷` I1H0!٦)./.9@36[`gq\铚w/!{Arg ]sQ灴NSklί]v_owq+(Q^]&>h.sg;Xۿ %z]Aӫ_?' y:PxX뚞\X<+,FӲn3hE_{;6zl{  8H^ RgQ\ /> #-sam$o ""ϭQ$2E#oaPJuy_wu0:poM=%SOBH9Gܮ\Ẓ ; ; 4^IcT`m@.%M*8cǘK!`%8V.+,fz唰1q>!q!20$y]7~<3Ot掝|~i]o~F,v`hf>_=bd: n0? URܟx5Xzs}+r)|FCPx0+aw؂}4,{k9ZZfC|M'(!p$F!&(ŨP9L KKQg4YNEA 9-?%&yZ0x@ \RKRnK* \.v#-ؘP݅P##eG}']zT_nؠV[T_McB;gA#5ֳzٯ-+/8vB eQ\URa&ڴw\'cB/a1!4gJ$ ˠ1;!+a7q)'e2]Lr&|?>te18v_;+ l꼬zxz86~R:> #G i%z;ͧ_o?*,jĈc!#a2g]>RAκkAw%̵uN^,SBJk+eJЈ`ǃZ!.X(1MJ "DG#q` rZZn,r Ur3+<Xiy!-JeH_>?>߲ԇ16Y1='5Aq#Br4h>}h 9h3Q#5"ʭQ<}AKqѳ-^A1}wo h<&1 G<Z^q֞=M g RR#-CQ =y'c%NOr pգ,z8v/1W>j۾Ƽu1*v>t4L>T8 FO w)ãFKUYw Z0VcydRk HWODː:BW.B޲FDCH׋e[fȖ%@a_ P">VcPro$Z3-vYGNI7x5V.\XIjdBʘNB8192$9bO})GTZA};uvD%ˣ롻Cb:A+ƳE$-Ȑ,0 U'#+y@TZ."@8w+{u&t(t"yiQŭ04|*IH> *$M0uK,1VV9F\Ңo іRd R2h:Pa׆C|nE GG ^=Ő2 XHGAܑ4 =H[M:O X]YF*R3 Kє:,&) |hZ1 :JpTX,!X1HC:2 FnQNp-$tb9Բ 6[IF5./Z^#)*R<3o|5 Ԉ[l8fM,ETeMsdSPRPQ%aH)/LH^(\ )DVq@GLz-ƹt"M qrjՅb"0jb)ZF'i)-n.e,lsJ r$+G4qjJ̤4 ͱN:FPEarZYW߾6[֦:@r+RP%% >9ImTDsk *J*2AJ^O ׂ~c{F|i$A Ǭ3Vq:D80byPXCbPGI+v (!pPX&*)""b, <Ǥf[.Tb4qVB% WGm4BIL4FeQQ)LMX[)!-]/]!(7%pWs/<@eߧA(NXщs&(ERGD%3J!-@2㛵\`M8a D 0bE2> 'djS0Ǎb\pJ'Sh1c:[l ڜ)8AvNvTKiகX 7}MZ5qxqVs85i_?Bz8?9!i) D"y:Ez 1;J0m U !9G8 8ezk*$X1$R*ܐ=cgףב2vc_/ܫ/(mtW//kvC9wC? WwzϽ|aM%0$RhCŝ*2MkgrT;6 Oyl`;8nK%8 A1TBGI |L+uZc~ خFlr403ilR*z!0T3P&%,P H$!5dȈ ) <*5!䐑i/8 ITGe:aKꧩc<ؙ}i@לּG\eJXJT7 ޜ N2&9MN[})p:X.9V]SBQ&IE9֎g\p^CQRţ4EnϜܻf`0CZJvcH_,~]㨷.0(rIhz'+NxzݭAo!w[Dyw*mSϝe`0& LPRz^xZ:\LN]hT1X O aTR/P^My)e Z8C7H@R Mh|ƐEI$cgBri/I@~tNrD x>~XϭrIJZYZjV\!xhՆk4p8. A7WA),I>) }IE"|ROI>)'E“I>)'EVOI>)'Eᓯ %.'E"|ROI>)"|r"F0rk>&~h筋H)#EzH"=RGJ:"=RGH)H"=RGHAEzH"=RE,qG|G`2Fu-@(%NQ!3nj. Ţ0M_H<8hO&!@9?J#89 QHU' nND)D Pn "d4 : T ˒V˃F[7XыUϿV{D Ir\x)iX^ 8{ nADOR-Mf<zտxazj* m~~T=rY8(Y&{ 8v  ;&'gJrN?arIzrm6-+y(WP(WW1; ]l>e+(WRq;Ce&~9wյmɤg6)DoޜO?9EsZlNVw#?n?!Κ6\qO >Cr@}4^.Sc|}{ᇛ䃷lIeSV{e\4wqv>[nvnM&;mfۋ@Ȫ'I ?jcpFW>fX~Łbo}`<ݴϼigjSYi#W=uU[j={0h>?>}3DP{JOh/?cm5JDž]4˕>Gsoӻ?O~x'ӯYqb8YU&]<?oy£UWM5lh`]z!a&NՀ^"Erc%bl8ŝMD5ySV\ ")qDSk$HQ4&ZXw163oZ62[Ϭ%*(Ren#_g*Ii>tH!;M[P]\K'8'r/ U D'.㔍c/1lY=9/cX8ϐ u2(H9w5X#„T6T..dudOyuwT0lV/FM'sxSk!;qFѬAtE?=5#{,Um{wt[ҒmCsz< A>u 3 t3 Я emlϳTŭOraŇ^BPzhSkap)Ϟаw&%5 k)?=8 }#k{Mdz7sN Y~>?j'[Pɕkp m4[Am!WļF\ʭgmp :v$?}R݇vېϽ#z[1ij0t?\׮D֒hD(HN1>L!Z gp<^4EKPKH!of" Q w9&Ⲫ~w=yUE3\K{δvt"Es[$mӢٴ 9bG7_])oqW!Б?tYr|߲!::J@CRr([Q~pA Fw5Ӛ`u& uƣaxYc-9ȅsa=76p9I@e$gSG'J*SA|E|޵+bBxrO=EQ<y:o3$F۵dʗ؎.LK1l'[^7\3{۸n[WwZKym7[Xǣ+LV~%a`.gc||ZAhϟW=h(`¤T5L]'E \C`. \S;.XA4zxQ^!W0%2R6* +OgwUDA 㝥u@@Zڨ^6ŬO~GLN;Je9qBK's&Coġԩ Y:nn: hR >F(1P;M YCpE7!J9+Ih(okWtx.WN޹cZ,]\~{awk6[=y)O{d_|^}Cm\r4hfŕS -Bʠ9BG)ۨzp*#S%j-q4KDHr\1t@ĩjRكMgY7騺IH5<)e6#[6='!vWZ‹H*uB0EvIDECy{s!ҐNmbQ$pRZLZzCWnӋn:O/>3|l?nǛ%YX!{]}_Ɨ!.yk>S^{]pr;n{{t'x{cI̐bfs/ ~3KyuLy ;6KoRpXuv(kɈg$oet__59s89zxzʎ5⅝lƴ{vVkRͧ{2/Ln#5-´uӻSMʑxp8tpl +]ONm>@JkSSv. P '@|F )] ^u޷fݮU <<&*|raoAЋ"hgOc/VDCx[wMU[o?ZIϥף~}%[WV"k>D78 =z}pĄWS1O삝|*uG{qvL8nFL8n5*hC k)LQ׊]&c\6wŘY#&jL }l,d9Ȝzr#UOƘݦbU\]}J>'չ=A}n;"}qo_)q;|`~܂7'3t&.@򿿟_ߟ] ϿIbINF?1iJE0\֫r l-.5>BsAm(Y`﷤:\\`u2^Yd}U*Xlo Ȱ\몢~KzF5ˇӳty|1(MaJŹU+!G*̳M;FC0@#SXĂ(ec,.j#+ :=w+]緈0{hCsAp.λY|ws a|?F10\Cks at*Ck# !50\C !6!5@Cks a!50\CksJ;7F70\Ck !5 !5}\1ca!5bs a!5F\Cks a!50(js a!50\CksO%|aQ?0\Cks a!̵6dˇp&+pw"=ҧiy2 ;$5uŢv؃NS5LΪ<nJD|s (yVږY8?-xY;|chDdNqƨFr ,I7fK}:b3Iqjo0+}B#[(Π/jgm֭t 8+JB}.?R#I%'&mf5ad](ưɿw>rs~,[P{[uoKrK۪:%cI1n&1®8VI\N S]TpgU&*qwǞ%w'uG݆&=r949J_|ŗgji!0%i'V(\JC0MݩΉgO?xݶzqޙz(JPd:i9sm9m͚ K6)?F?ex*l#iX9&ob%Lϙ3ĂN9[q*Dg)jL*Z7/ɕL!X& iϐ9sm:]|\] ,.qdRйT=y tvu}xn}(imBZu{4&՝Ve[cwOo'?o z5)/Kb1\bXN*sb$wkeoPooi3fݞr}j[GR6v܋1qrL0qBB=9$8`XmP\=&?5lJb"$UE=RLۨUI.r U<.){jO=kK<\VztX$ΈtwV:6Sj#Gw4+ or{9oȢ4y[/ߨmw7zJbW (EPuY1!1\s)Is9ye]cN֞WAx4J7bMV۠@F $S΢؜/}qc@/\4D%OT@b)U{.VQtg+7p[-_!ۜ*bHBilq:G l+J;7 B0Kqt.<M~B~eRIE Z,GM8dT6ٰv"&+L^`W"ZݠVZr%\|ũ|'=:iovs'{>*L '{;Mnwv/6YU,o}_tS)JT)&30 &Г0i ،Rc O)"4 ,l =>Cj 5QVe]1ZH^h Mhc$SwYƘ A4&j*6'EN~fhICR{ {ȕb*DT\>gZ$0N1aqN%6RF^h-~ 9cFLB9m: oxZ~?KY,ـ4j1d@RdϬu` Z{P7&k$w*pC8eI{mv`\Ze6 #+w@ZC@p9%&8l[D*$ez8×8$EE-N_S_UWWEkN08Ѡ=`! N 0Q 'NvF vp]ok6Cu]k{w̝1G#Cs_!>9Tog"Ä@R*KLSҧ랭V:さ؎;p.ݧI$C8 L %qUiIuyhNTŴ6*8c.)q(z 3Qʘ|t:amν g9AG?+.Tf.Ms=8ly-+=P: ([R7`nW } (&A&xΝı$sb|y5 v:,$&%R%:„T*M hE!J&2(.w<=g9x<63>GAeZp ^`tÛ4YC(5TOB}`ڹ*$E}$ aTEV4<*31zFꖑJqsŒwA&5\dh!0K,!$Cs.͡(T=QԕX:㻋2) c1rssO|^h!)`f.. b;qIRg}?(3_W0)5O] ͓璀B5_L iz wXQ2T@h0SS _\+aN& ?]L .РY69PVx8: Y^Ti>!|w>!e|$;c#y`hs_ȞKrAÿ&ǪN1[:@:Dž9폿~x{B7?ZoA1\M/?h} o~}ЪafhCs g\[}>B.(2h~[+b]`2 oRS|زК`2 Eh6|.YdV, р!mѵo)j-+ViMF E\./KAo{wgy[Dɭ6@<|%X2_^4-B^FyqQ 6%6{o-jV։a'alPVul6%qrStnuK%( 'A ~^䛐-sw-C=4|Nf+WMn )EJS$&߈ q5r$&T.8A`tG]ű'xK3:JݏG? ߗӡd + RrU\$#څh;T&Bh5D,=&l[;.8zbJD:% ȩP#Q'+?dHV4^IcT`m@.%M*x$jK!`%8V.g,u<=tpپF~&>V9ܢ?Υ+e^pJ(QIym 2Q 1QxZ˃0ĥ X,5h9qoԉ*.\)N Ù!jNIZӟvr 2q);àln.Z!1w/e): GTb&F'cb0T@ҹd%IPꙷSY"ia;Q ]t6޾9}ñмoajQ1epm .\&TEQ*K >Yt} i^aRV5TL@br;j˪\n*yV^Nno*}D)fUw_Q{Ŭ6/|g0.'3꿂7Rdf^$ڮUX\(~b4zϧ0ƸV7/?~W/ HP~!0j#>,MfvPro$Z3-vYGNI7}*3پu:y]5K0!eLXό'!\cr/Xs):/URah,rX q\e ;-2=M]F7 l` V>ӏJUWP cu߶n}o܉ oms"++x1Yt66 iGD88 JQ>Fu.1bYh3/1 \1$TTdೠBSg Z%Bsm-֣!Og$4Z2*ےLd#/ӵ_$"z|DQCL`9> H<%1hĝ{tz|;<W+qfKv8v+RTq4$EY+1&AGJ%3+RZr|H+JqN0䤎EvcCBq s%*(ADI3eA4 Y>"V+~gV(tpuQׁŒE>*>zQ Qn/./XxI JIJgUt!/Mc_f|aրn1sB%F(/R|) GslC(1}{RlE B %#*q2!^l ly%M۰OOC%{mfU5ںd!ZN s;(6mITDr:Qq !Һ}ݵ)Uj5/c~aBPӸA $ڈ1R"َb<L0&*4]`c=m w@6ABR+M # hPl2"XTRۓc;nl?}*t`;Ճy-w >=N6]`^z 6)0}s 6'TAhBC=&רk|9aKT (23V˯_hu%HXos[B!`{̒o((SrWr!?kȏ‹!?L]d0"Y%FU'k$L Bu]eKm?6`@<&%e:O7`ڐͤw1 d/r-fّX7U捞ox-Z |Y B@"kSZ) JdذWC ԪY cZy)o:&;wU .%x&%x@素<:)kVT p >1/^XpOhRGĆ}}/rg|'(Kİ0ɈXF'GJ6?Vp P"?aiziE |%Y\7hs1ʤ" +{+B O1\$Bgo %\UQޘPi6) cMkϵX-h#Gl.Ի:٪'' $,KO C*cH_oa6?VWzKa|ox`CލEMr8~/#8d:eBj>g:ٌh6g#x~Dc-EoXJQ{%S W8׿Ƿ/??^*|߾x/ GW_5b]UaA|үm>^KzOhzmHw^~at6Kujz͖UYY]\O(f5?rH&QtoT_t{!|P;rhxͦȢejRRf".gmMB.; Y+1-mQ|T%UbAs:eC&"]H"٨SBR.١^F,3b2eS<͆ 2de)&;l(D(^O09IZCBY>&KƖ,O,g/4lheuf`G.IךB2HUXr ټc U `s ߈;3붸g8t'-ZFy@ymLgRcѨ{,{NOuHe i)M$<NEۺ} /dNX~F7jRƶ"wK18vPҰ~yMeطYN,G^»7. m6MPkpv?T`=a;zw{{<#yY>}Kiݽm{oFGZY{<\L[[]~]d~BAs# @Ӓ<3OSoEs[w"mH3O{3#ERm/5jsKYEQt5֧Θyi@ݵ Jb L C-Q4u 8TU>>5?EI) fBe:Fc N Ikim1::h4'F2 $dEGrV`9A*h3A21j8ӻHj-m932!겢s"4C=@q@G/ p2)f_Q86'4>gZDk@ic[Ht5Ol(9)7 @FbsOu#zt7ZF槣I`-mK^|=-QY:E}i[ RFw_4Zv9ν4OGNi2}u+B>h4m w:>6Żdi!Vćxxr6ݣTj<-ܸ՝{j oEzEkwZn8uc/z!ɐ+l:YeɨM( N{uB&Ȏ$]4wEZNEG8o uW"gMW:olpG=M^}y>][Inja`+[ڼ-z3+4 f9f 4BExo88Q1*Td61Q=n$ẏ tΒDhG3q aܯ*Gyr0'sd%4Lqv dF>ShƅrK!=Jr0xxz*[EX^>%$|f [KV.P?|IV-Y͎NN>qpJvц蠊9*=w/'g 8?lR+׮{ru(EOd\V"y 4LQ}z{X?+NbL`U":R\Ƚx㽑`bof]Cn?Y@3k '*_h<ӃoywPYyU%;$7'wo*ufM8StNh A~˛rZdv 9BU3%ַ>Z%m.l Kd Bq 㝷Ёq Oے_,m"oV6n?qq$30F2Hf #a$3VCjX vbGCCj0!`5VCj00-)O}) ݇+#HBLV(≉ͤ|eC~mܶ˛z\K`J9mulUu[+h_zJu !jQl5)=)6Fo⊸}w^^q\l͞'ǩjژ"(%-0ڻ(WU:*HN4i:\,Ëۅin0H3Fߊ|<| ܤ~]X#IhyȓGđs9Kʥw˹߼+A M>x4(˘k0%2R6+}7owxy:o BO7zovo,u6lΧV| {7;>AsG1ǠD1DP Uso˓ʐOz#NU{䉱ԩIu nA?~vL|LJcd.5xo*fl-לJޤ`5dPLEy[ϝ X:"N&qؤ7dM%7wGxM"crN`ut2 PB̔ %Ly O딎ݿU5 ;x>v Fk尡I̹xP!(U ﺶ C2o+8Mjg?zFޑ!?AL_́W?'cYrl׹`&^-_}"pL/ o]|ʫjtv|s1=_ gm9$|vv<7J/ `cFtj'> 5l&7 7ջK33U(gY*s"d袄<>+8;9߯nhǔ˽+p>aOO&f>~uOs.PN+\щ'Z-S8Vj+y]U4x+|wޮf)v&{7XH~;goq?M/5-NϧC`C.҄,OKQZPݛJE۫>uF=ꌲ MZch<9845qp…BEd5#C}c>,J񮚠u,%$U8jj[$6:wݦ@ƍ eZᣛϑ|0f@F`@߅o> I!8m}N~"^Wv(AT+u1W; 8 l=̈*Ύm:;jy?[ZZfff-8/dD|윈.=-f_swCɪ\E.6XV :|'FlHie{[|,Qo("6.Yb#-o7`+mhĶJܴrČZXD_S*a tGl>9B̽a㥗Ilr=MYș''0UĒ.`c&B&k!ۮxXGj}$Bfd #(% e$.Hvx /7"0 "vf##-DY 9ip Do1ʁЦLΠp!oRAb⚈{Ԇ395m5RHZ\jhлtȣfF>uvmqQwE=pqڔP ' !Q~rQ8/DɓDDۀ.xZ.[2v.pqo"^ֱ-x  1; %zsE2Ip#yRAُo)ˣ2!Čifs`.ubI@MJGαqٌD$BVۭE"dqH*Bf  ^e`j$s1&ֻhbO P}͏z ϑkx8y&. I\#~L.9EVnYUL`ZpPܹټIX+p柨BW5"bW㕱cWa х ނ&Կ$vj)Vc1u-aypꡆCjl`J-︭E*sKĥwZlԮ5~2uJN9zD,lE4!<!IEg ~ E XD,%?dyE9YPwuJ+(tuyxm [Qdg&-ehET-hM(^8Vo@om;Dq]7y]+lA$8Fr( JSat ̈́RH#CNܡj"0 W=`H(pVB߫Vgu)Jf£L%)-R0'Z=K.iBL>R"Ӛzx鸷RPX5!R:(ꭰ:«c*]np ng_ƾ*ː%00A(l)1LG<\OW4r^ӜG%!tUC(4|&K6{X+7I;hg*HQ'5(p5g- (GYohޢ f @7 Cdh3tI8vgj`sJj .ܿ%BWE?ߴm 1]њ{b;|@ձ B"b$}$F[k#0S&*"ݸ`a}0Iju 6JXϻ>sv2{GUx˥zΡ&Cf =;a>6lbsvzw8M9]&RZv#C= vt2_uwn;:6Ok(FSDQݤ6_3qEM$:ϭOF)? i+b\PXZ*S*O} JV[tSb*JyTB@lq *G2ʄc&JN'b*e^RuUfeMRLN%.:$kRjԦ8!^%'9J KLjh:.x5YV5-=yU/q<.‘)%x3ʉP6 ъR-*B\d P .0W̜i<Dd%ZQP/K B[4%b(iZIĿӆFL%$EC|4! TR IJDrLHeAЂD)DIN2]!IM< \I$:dXN{; 'Ek{|q6[L sUǎf6rt!ػrCh9p@3)N#Z?2ee]XE=̚.p3_b. /t@ۆ kЋ8&,}f7pD5:9(C%vtANrdgxξhϏiZ䍣^It6#yt}1X 7| ]s|>KQ)V 14O]C\ΐ,g ge%zx~~qQ)#׊\+ ouh%+(Iw3-[79va/]xy9?]~j2ԕԞϣq:.斛] ̆||ԎW{?!d;Nnöކ361Q8yJ>}'zvuh8S[۞֫r?:o,P|~L}r+59|Lh3YDٱSA͎N%FƏ.NQ޾_^߾|_oS߿WW ̣Džh|pܸVno5qk`%z9W|->B^IzЎjmwo_Oقd߳R]rjW<ʠ+qhB6y-7F̕3!1À6v ?V]KGDE$rh:(RwI>@>!9315`m IɑJB;Hd29JЭ R:WQ{m$}":xSL&$6\=*&ԝ'ѰMtԝlzfa{vܸG͛NH8n;dӿ8ӏx7@Ax.q:[WG`x7a6 mI"Z7Y|q6~lg06/Wd4~3wMe̳\t;xx՛_qr $;s8zjQd--=O?BXj2e*јZļN[G$*r*r*r}VU.Y%\%\%\%\%\%\%\%\%\%\%\%\2.Z;xjZ;NNiVkUejZ;NjZ;NSUkivZVkivZVkivZVkivړŠKY?h3.ao8;WVDu @D뢶 h8euuk${3S\{(aA;^ZE9yd SRDwXQ4h\kJ=!&ji.*`$с>i*Dfvbù?n܁zu`Ly|pp>bp+!vݹ8VtDPQٓ1N1 -¡8s0#R&sZF"F eg"G #|sވ~8lUb-fͧw R^4w n^ro(4ڜŻca<4•cx݂'7`#ߟ0q۳ }* ggӲpJւWvZ[h7Ŷ"AZ`H v69=꼐G"uHiOyRuA KH!ipML09uE])Xpy瀵lD/IL>XD"hu:%* *7F+BZeM`5i;>/=ʛ[ׇӶX+{"G?xwӃ>R1a ?'#>G6Uvo/_Tӡd+R* bT-bbe80(3Y`jDzM%o#_M!<l)KE7Yy_gӶn5/ZLqm$ aR1)#knu {Vr|%y

    `}YIqVuJ"p%c2P0FBGruO"ZX,0T 7[H4x#e`&'Ib$ZM G-x]+4 a >P1@C]fgBm#_ e" Yo<+l gFY9~:ӏ3trݖl< (^`C?ny6_?-*9\neF!*G$2Ai.;BڨU{C xEjyXt~*jz,91 !6#uQ8Sv=.ѮأA?/ݻcU"՘T`խv@Lh(J).jZC2ZvKP=p,6%(I{25ՕؿuffO,[쮺6o $t@.t%~41p NUuώ>--%ƞsQ7"Vʠu u*]|RI$IVG;T%kEAZ\3yj }2ĝ6 P+xKvĸ~nN_7X>M3@k0ߔ W_ڂ_Eɧ6zv~hF\n7apuL(Z`m~v\wU^Z,{5roj@p+k@g)] gyKyWϧ_Tb\L6 /lLuAW%&>[K\{/zKBlelyqL}X+sQcm !E:ߝrk1t3RbKxiĔ˳|!wf_3r r d2Ov:.Arɓ ؁je9{ނwfu'mmoL;v7R'7]߽x>l#{.tãc}LC)n>8ToۿS̋ռ[n$a8 :vydTȋT+SJ{UGBkwom~ֻjfm9e/'ǥchebPާ5v_;?+< [}nw>݋򇃯O׮/9~>,21:`?S}K urdi__' l ͹M"GAjyujY8|@) RbCWo=>RZ8RbZjB dsƢEF>(銳x oYgf% bĄ4-1;mOzi3l/y ~F_7ַn_HE \ JGe>KᘳFRx * ikuD_RXU"[fNo7:hrNٷF:6ڬ}Of R7tE-GLiNv PQM3Dfd9"GJ!\2PפcW2I6ОO 0{O%1d6=8 y jkǻ3Sb1b!x3K[}Xf}JZMІҖ$ŗX~t*qcT<@#粟ϲ)~MMx.wC6)UߊD砦9fO 45dp 8ڎNČ(kL-%e74J(6W|="NqЃAWQֶ/{{0&HIԈ][D/.Sǥo})4[/(^!K؉HIP6B\T{Fwލը/i{p~̓^~B9Z%u[; gn͚vڨč`6m0 , ABV>Ԙ )K=fKu0ϦM~o.0|AͤK1ҕQGcWFZL4`= 0LMnӻbq,45,TCqW"tΘAG]LYhAiP-xҕneVnFhQ: sAOG_1뿺ACЈE!o 5Ih!`cW:'BlF4st uyۼۧӶ m/|LΗ';eCP13Ͳ3}XD us{s+0_;Y#\ 3n~: kw}5F6d jy]eK?LHX&/lRv_íOߟGi;o7mګRV1kFĕ2m~Orsqy|GGfl-$\5FHP6<֘cL)d0$ UtEz@gǘTKn}moYJCdh3hs9:7l/񣦝h9<-Y6,bFuW>ԽH .NϏ8@ ޱ ܰ?́RY?e!FiʾkI( ŝ죐΋B/4B'Շ@Y1׹G ܚ{㒚 V!h-B pKLGd8m:dow7>z?_O菰<ǥ"CՐPp73A1̋4tQ!\(c oƮ|jN>PH !*UevLviyYo;Qt闃:f'b~W~hr~epnɜ[_Bf]wZ'gÿ_- SY%ٟJ9oH&h|ym:˵(esEc^kWulHzZJwn5B @(}|ɑqtʴX82 XV,|Xx5? d'>g_ʅ.y{?svE84DJ Z7Dx!{T1=ܴϐBG{ up#l%Fs.C)woasuW8͖ɝ5#k`,xlΓv^Q{!ػ,~X9( /8bz >`y\HcB+D4NeC6dh,kj‘ͧ 8ZDkꊑ2;wt;R\oCQG\gL*cY &coU1@$Jl.U 1 ƍa#C>V- [l8T1is XMZoUų¹is<6.+.x'7 6uDsv#-F1!EJQH r!ڛ|Xq)0Yx'Gj֍|CgY\TRU + kJi5gU{@"ǏϹP)¶~\B?|DXטāJmXKVG $֒|M/EjhĪԊܺP`胃's#7ݼ|)NN|>4 Tyx5L9e{ [?oQgu<o/aYvaeqיܕ]bD+zQC:H'Zqߓhϝuv v<8''ΐg}b~^NK^g,y%uΒY:K^g,y%H:?IÒYK^g,y%uΒY:K^g묒%uΒY:K^g,y%Y:K^g묈uΒY:+..y%uΒY:K^d?2%!ΒY:f,y%uΒY:K^瀹kx:w5P\PR*jP6,$ 6- w>alO]oEE3ʘڐn$ũ@PYC!iq!S"f-Sc3"RP5Bڒf^0>BE!b(yME:CH`;[_w<1FOG|hN>}R+k\6l!(Ce^~FcQCcnЊ|"%?r'I qQז[$$Y͔}i~nϽ8$Sg AQw٧Ɯ]N ]5[$~6m:_}}n3ԓ2T+7/"n(<4ht)8-zaݗ |~+_I\1@ck$ Y⪯^ Dh(h,5 jœOrdVnFhQ+EU3 (KNѭ1'p,xBW~e:4b{rA[CM{H>غGꇦ P2]B]6oGm<ؾטf"u<.+b 3}XD us{s+0vPJ\,Anw7A?Bl$o@Iz$F*eBJ6yd9d]z}&4*]r[rʛa 6@ԯTЖEw8(gJ/2i8b=S@$j<@ޕ6#"6UO%%tJgVIqT,dp$!= _+1Hz',(;cĨy&"e,T&$0!x&3IfzFꖑf8 ,ydU,h2UA:SƘ(tF1MLVI: Fa ~]kii3dA^ǙAapWZ\`v`0N9I'x* 3Ln=zWlg㡩a񤱹dLoJoda@Z^U%Ф}&t[RQ θd얝]-%x'tM$oҵZMo2齲|bNCfl>]#,҆ cc/>*JLsa;P_[.eۼF{w>=_QZtI{!?\JSzqy6 $/w_~_+(-&i .B&6irӁ.Y`V 0!ڼk,P{֚t;uu4$= ZC?dE]`h^{M|qϹy6.N;[`UsugR_ԴCl f"M/щ.xd`\f2T*e|&L`-J:vwq|M@ >9n[ i=#9tE=/ Vsdu%=2$~\F0[ PVK5ن2^W/4XtwҠ$/S=Rr@~! ߦMn }wo{w}[M6n%,)ǒ%X{Mj[{zh,խ7~}=-k[ r6֑!-b:-PVsm6֕U E.vWC B IG]o u=$WYkZ>\70zeLҘxO9skwk%FcnC{0!T%4֗g%eyT|V'|r,ͼp"w ]#DNܕ*U)WҔ=;PP{Vgl:+5<^vUI*4eL)9{p*48GQjL2%uB"51 !Hc%Aw[Rc2<'cqGKL$ Y(IRdܣu6҅-?#qLvO1T7a)7~1D\}S+w1/aLB6.4VYq2ȌdNzP E2dtkNϼ0%fFi8NMK)8k(vMقo{H 8&᭿iIi_ҸZ]vD<ӕVVYbYk5j쏰I'帋 T [OQtץMqm%a0B=*7bA[˴;OƖsw );{a}E_:ߗ1~HG,Bg' )IK6/:Πb%#ӹy2M U7q'QSf"6YA+'RVUvY41[D1zš\fx< $5 }2&%>XPwIyiuPS#'b bZΡ{K:b{s䱽Eܙ*sMSXr酠 nGwbrf%y$)\V1f\THXdVcULX1 r`Kx(딃q 8'OMNxscP'|t8 qR/Z]uYv,YJ&ikOuRnbռۛ,#qN_ҦrJIR#i_ݱ'{,羀۫{Y*cT%M6ԛl6?ON u,壶?s|9^0,֠oFk%/Rـ Ȣ&@ϵ p Si)- !IHL,KHB˛iz  !N/$fRbȽeD!L_4"XT:IXb%v @i?+4+4!$j)Fݵd'.Zf2#zZx`W+x79B7<B/Lc(5DiCo1 xz{*zz$'”VsA0/D;kkc{Ӊ)puX^=G-/D2$&#'rNquq;,;ϥ "`qagf]-n \&4ox)a=°pc'ojXv,>174]ph86]A7Dh d%:N>JZ{} ! 4889{M5;2upA<g}zH"c@_j4*JeTiF3j.Kкd^'BP jw)ZmK"m/&_Kt O^Vrᵡ(=}<ב⹄v{tu}5jEJmo+4X0Ҙ%89?{ȍ= mY X̙dOl.eHȒ#s }ݒ,_Z%;=HlXlUůuS+yJ8t2ҩH rDJM-7&^*ֈĸN#*jCN9)u 9%kt!9yRGBT^KN~'CΦ`-)z_\.zn1֛7ҹPlo2A?{ ϟE AMsn bzz 4U܈C6-lUO~f)Jk"[U娋.Z1*$*%YժBuOxhh)dHg˭B^BywsV=_ue4O>*Y<Ըe+ZCa͋R}bQlQC@*,[ApTd~ f~}D$~I&L6q<#up՗>i1NgZL3BGl8@2*ќ=`a?^c̬F3PƤ3Kiu9 3f.ZY$j\>/G]-H|hzs~8fK;xp`G2( *+ovz2+ITK]b`!T",_O FYY% T\G# SH16))v۪.wI/Yۿ.^#[9)ОW\)d :8zZT"OQ ѽekըӶف%iޡ[@紟V"ԨhJͼQG}J;QGesi7uh[$. ,$@zG0Qg%gd2N%ytE6n (5m-)@cQ*Kd*xN&&„Hb"GǢ옮l%Mo%Wv å.?TP[2E2Je{9Cԗw_V]Qxl^` $l6Kmx" "eֽ]zI)&Hs༓@ 9#Dh#t M R᪠՞YYӄI:K-E010@˓ [4xL&,̲f&,1_1 {Lsq *2K8kBG2QTD _7\mUWu,V63tF f# fXJ=lɻ3N3݁^j_\Uo`$DfcEzrRJ{! ,AXBRSQ!Oy^X,L%Qbdxv%Q-3B7A ctلnS-a.HA %-hp\)BUr>҃#aZl:;j`l~A$ϧ_kIk,DMllZ"oᖄ4M:CVͯ--]/Nݻ\>vҭIxnMrԚuevե/@h{0;i ۜSi<]5OwiAy'Eֽ]v.-{^{r3=Z-['vZ4?,)WWt(%ygV"jEj5SU`>f9J6ҡtseA~f /|ώ\eP kEUIqYzW*J_ :cIZd2)ANdLʘdWL@M+Uf%MܚѮ_ѡ@% nQ2ƙm*(E@apt5#q_!MD:Vˌt 84:[pUf{ֵC̙4NKvup?BX¤DyZALf!+,Z$' "PBcڱ9*rHGHNa<ے8DJwM4$0gT@$[Fz+J0k(<2ťu,h*]A:0#9ĘpN]Q;:7Zc$ϿZY{ |ؘWbB+4L>U)i"I2Cfӟl7wM ˖RG$6MO2~p+Ŀ}f%Jp eeK&N.ə5Rr'~6Kad6'!xqFZ-2齲|bCf!?]+ZM7~ 9>P(ux7W󷉏).B0 9.}-HD;E$z:Jr"G R8wz\wi)IuiΊs:B5)GkZ<=h/56Sinx{5?_|n3+5^/q Λzo $Dj㟯uw~Qkə֑aD0=wub1QeI>.7zv=b8߰8)l먂muȶQJ`5ӒG篓iy5#|O`XoɱS݆N ?Άys"?o~?P~ÿ~Ǐ.û}ҩ'ځy’ڤm A9H\Il_wh3cQ npXߏv10X#7 4K#Ͻd^H*]$&P!&@p0U h^ ﱷ;#ziXջuhe1~^3c A}H3JX/7pKz }2jW`5F-!fJyǓ .sZeA9ƍ&% ̱O=í cFdt;sX_'?e~,ɏ۶xzzYp^f!NSB]ӎR X!E$z]䅓Sov,UzMWa8ۛzy}z\_[hg.Iu"Ÿr4W(JvElt{eFݧ:BJڢ1A`2m)*^)>Qp>HOfU WF`b`ֺow>۞zuF}{2tzx (sq<;_H!h}ޅ[:R+a҉0]V"@G<QdwE$&btHyDFgGE3p&Gѳ8`s;x:IBW7$\~(QmUB3װxD֍y.! r!B91<[/LptQJ VF*ؔLKD4@:&A @yұN99@$9}Cr=xC! skh'|ts8q)C2b^,@u$b^k, 2kOzg͑;Ȣ =2{\VN;YR0I*/L=R] {̒C$/jכM)B8FMXfK9;d[F(&H0}4K9Jx7CoW11egf >Rz>*-ϧJ~*ECRx"l BlI),KHB˷ (NH2BLb6,%c6 d*2$?(; ܪ, ƨl,X6]V mKC_.U(EokΦCvC/{+y+4#$՗ZAOlp{Z aQO-J!0+<5ho=)y.4%Z0'D c#X nd:?i=ymAcJZ0G%"FT;HGcPSQ[Z#8N,9^</D2$ 'rNF1tpq;R-s}w\륢{YoBt 65a>Az&^B^Cv*1D_TDMXTh0 ɮr@d8ã+kbr3ZV^D6&E}qSak;V Kmz(®{zW粨=qo٠nE|)1>x`D¥.,i9T)8%# A R"IQJ*I$SAL5YG)!tj q]%im+ IbtQLAiLޅ0(C%rrvIx`?l(+9v/CeT Nd%GE꒤/3u efّqe{X#^Hs@2dM.N oY:(I_.%] ȱU+Xo_){%cJBJ 6[vEg}l|\}J"~397'Ӿ/Ӧd^6Lź:'ZC i1*o|nG_)BF+F2Nza6*dWEFp))LO2v6JJ%7MO5oRqA%--^HD0Rֺy{U:DBYcTJj(aԁ 7$}߭.:{ pKID%*tEgJEDZv.i m%{)hR,:183IRt* rPuJqtS{XА> !32PKu&5)NYh8o &٩[1mw0"6,~Ϸh%SLVD: )ֻr .͆IzJ& IAچ(èpmHAD2d|*جuv!p N2Cy*DDf١Hm`Қ92y9/5yu6_Gf^Q~4= ښY\~Gkdr]Pjô~|z).'s,0Beѱ>RAokEtDBI%e-YoJes򺷭ǐ4X5QN,9/c(XKDȘJeR8!YP I)&FUЂ[ Y tc=k׫?5[4 .*@QFXe+a$%efcFX-$FaƔc>Tӏ"sC~eG/" G%6JU`>8IGTRAf&=ðyV3o&4IK6J"%|J0E~ĬBs R嵔 B=x0Y#Fp h3+ռ%#]i!*gLJZYtx_VjG,-T2(,QEDxm׹ïLDU>XڻX̼Z3BUA+`-eZ%\8 գz,ؠIȞځvxAYX/$!]?[ԓ by\6)5e0&R1@6\ڕ\^>J7~*%jB]HP UØl}Gyn>ʞky(23g VXX *.J)=:cj=qXpަ( [|"u$ҧAV ?dYFi @%0bHVaCN@rB;6CA+dJ&+wQAܽuNWKtݽ5m)@}./( ttf_Y'@u)~8=wl' ϣސ/(~KҙXOu5ҙ f﹍ML{P&Uu1Fm0Qq2ue쀬gǽƄ!aľC$ #k\kG)k 2h)8!91l|t% |f5Kw"NTJ$ t.B۠5fBGDNDj/l9;Kkå|tgk^˃-o:F@hX֦R7J@f TEY F4B JފWddf|m/'-6)&yV *}Si(FgIJbȞ!Uv A,[(|Wþr8zWOE߉̰yal(#P@YiתR#hSV*?2R[.)!\-d'b0 :$(~8 d,ʤV L"gAdHbH*"Vh70ZڔSV@cv Õ3<$]%9ubhQ="j$,]~y/׹̖/e >~2̮|}Wc}jX\c./~^Tif53NegeBjNUXOl/속In~m56%D$WFwQO/r~ }R7w!ҧ4#T+S'zK:-Y%a*IiboM ]vCmzJe Gi>nx 0{_oo>̿*8{b1^j/?ϧer~jwlW$,j|z?]>I?{ƭNKhj9f&a+j4eRDN\})H@ʖlpn|_/ Jm)#}MØX7sUX~EezQSl+Ճ]99̷ܜ5aFmuuFt /F5~YH؉|rҟfbA郓s5;qE?/6>ǿǿS?~?^|_~UoeanI 1 z -㭆Lϻ^>{qo󗒾 XX[ dɌ[;Vi]rj +44Wdڢ v:xyŬ~1/  A~jPOu mHhA< vºo'@WRh1XM--2ءB@r+KW.=CY:LP Q%MVlU,&{vzqT"%.=:]u*LiyIn"qV;/ra7;eٻM:U@P0] )]2ck ߃ڟ.D0V}&G1VaީTBD:ϮjJ 5QVes$K#saMǝ"rЩR#.{;#Ga% *3ܗ 9",}LJk)ZTU:œ1#eNȺM 2짲d!6lNJ9) rZ`ԁ}k4)|Mʲ e9+~1f}}}քEFGؕ"pZ䵰-ةm$k2`!L_h2tӓ8 Ⱦ9ވ;3,Zoqόqi~.ݬͲ*Kx>cL-:6$ C+dp76~n l %^ԋUnve2piXZޮJ-y9Z]N$W\Wh >&-=6F6Gm ^,mdGyMsĔ,ֱP޶ldkERM%;sL{䕋K(e@ڑ¼1VaNan:ЂZ`yhl W QrC*HCXmFFs /WL]aWcuJ4);fV:(erkgO lPl} )\U`YMWlK$5+HEiզȿ8Řl3W9C.~nӹ?`G+,W<#*",9ݐF8tC㈎^Y0  N_!1e>)DGbU6 @ Fubf6aDFO@F*ʡhޚ ]|1N(k)}CUmjnp`_}(WȪ|6>1W:Lr^[1ȊboC^XA=MsqzeaKa.\Y_N|]ٗ?ow+ ູ؎H념 xZӣO.{}8ajt*ɥ8TLb0)qmW[z X$[L.x aL3":]vPmFJ{gc`kVuSڼ&}gw 2֟?k=(6L֎ [>+^5ߩwҒZz|QP,iP*([M-nӹ{U,ffBRTlΆ3aZ]X#ϛc/14& K̲Qep۵ϗqgx a &):Ty!It#4V{ԨZ h1y c ޛvnƩ{h5dP!֮sgI//<61OVwgIܼQ  q9W>'r|7g{)-˹iSrdQeVS2:eP8@NQFU9jW&DŮbDDŽk*:Udޒ(UeE5t[q֫n*irqG)&o#|fi镦 \jzJ&GuI8L˚#xۻFW؃CWVsYUvxUϠ?oSҜ`PCm=o57TZj zͻ͋E ~MxuLt:o;>8Y0u.6^|yR#83>1^!e1Bd  cH`IaUa'/a 3%:st PBezUbMT[8T$u5 :@sP:4K@u!XB+'ˈޥVTxý#"m֭WUp ٲEЇ,GB Osr6GӳY-HR5e0ətR @9*`@ hyX׉^ˏM -ǫz-@֦"b ؓsD1"vN*軮-I_.j)w$ Nsr>A\C2>i=abS1%0F'l_Ӽ}Nν߀εam*gmp$vhϨP䜰cαZU^$0+OҴlHG\j3 kDiF~ϘB\Fs-l~y]:aMC >%W%'$ȮR➗H{T$HU%6:Hg_cM ]Z`c S-5J&f)kǑ5NƺL1e p9BO@>:JQt6Tǧ+v~֝Wo,msA?7}Dz=@+n沾 ČxuֻH6;tep΃0Nzm98) 3Uy\!7ٷ&\4LdM{1:~llqgKkr6MrmXJGHB \Tzgj?R(U+6{@qBvH2AU#zқwҵ|fpίl!ւJM k˥!!p . >mcDq+6~eGv X]`3'?O|1"-\`5,:щ&+*X-ShVjSuUPDŽ>ʻDooAwny7R7B|[ߴ1Nbט-v7/oDevo.LNMzfs)gM݇&j6 H!Q23yV j^f5lb~<ޱi5K*ymbMֱԜ8T]el2ץu+O)<!m7 茿| hd}^x51A].9&B>3`o'czb=Gf h/QPoBW+ZsWA %Sq2[(UeS8y%VabtFFǭ1qk{ @ΞZPkM%)*Q sdZIfݦs7l6to5ys˔eReN6U;/|h!ЛZ/?eEoZbBmQ}c*DECu bTs%lkEIDpQznwDM ]5LXk 2NE&mP*QdΗGEFL[@zfSŠRSgݦsO=[7fr[/]WC<5k PT@W<&B4PiY0Q0%'WoEvߑb1hV%de*liѓ-hv V7'dI[zsqPW6a̰d#1bW^*Z2)91vƁ*y&`cKƚ0uT!GZfuCϸ z( k:9Q-NHWClN&N ] U'4Y-d%l/m9NVj;?(GSD - 8d5>V”DʂӆK bj]Aӝ2nr=*-Ak42sJ6j+L@&Xɣ_+Ggbڲ5'> +݊.=]LfË́i0 Nדt̗EE2 o<LGasLDςr7bTj f<8}IJ؝Ƞ& <$و)Q ^X)iIܥǀ;T"\1o7Ń2y֜^ B=ֿ?zqsde6$Ksm+ԒAt%*O8=Ĉv偤u, 5Y+8qOFbKō[ T6j9,|Ƙ"Z,:6O͐k+g=6$t9oft>qhN.Sgtps86{b3;.䨊/{ImZgk &82Cz$@R. \v2B2b /"c3"L@,zR( m΋)B mkt22VkX^kfP a“bEo˛Œ;W};,_@n2~Ěo, L0o8788OJmFh`:6 =oNp]S̊^C{m$,EMil0pfv \ʙRfBX2b4>afDZ*P{`f|:L< #3OMA1頹(q˪xX"L̐+RA&KX# Cđ0~>fY Q,֜XhQ農^Dk[,ZP|8(&mF#yVZXpQɪ"ΘqJrZ'r}L2!&Sh U54S..qi,9ye\.WIw6ǒ6!d!b0ӭJI'4!!\<.F9ך⡨G*9(t\|^/]# c:֍uy< n~|'Gޏq8k\Kjr P@0}8S aPW긅>eq=ʾz(a"Vv"e E3HNd)q.f%'y34B7;ݧ?pKPEw2Eiykf\:iZ2`O>NSZ+՛E5xj6mptIJ]l!iDlM4 CrkH$ʁvө֜-Y|6[0ug־xd܆=bP,cUWI:I`1ʢ(!7:Qi'CPxAG~N]ߎ1+Z-U[ˋʹ]uONnM]fyE8kɇzuC%mI8;$yB)LN/MR x\A7$tI>&c?rSn|d YjyGqTZ,N[t%-Pl#mOK^($䭉VHpD(soi5`TPYDy N[ =0a)Q$I"庿[ 1*2< ḋJ8 dwR Rd$zj{xTk0<gκ!]3ꞯ*b- tjzџmLɁGs=KU\h΍ћLߡ &96dFzbt 98{gV%囥I+}"!!YAɔd#wu P:V_60]jaġ*RT[ٍ d8R /.|,\O1V1e2$`Oȸ <3,KKy2dB0x-rF ?\<>f΍~~~}|拳Ǣ8Me 8pU\gR#:']+}dڍ|vOU{WT/m?Z־?8yV*u[05yʓTQSrN㶜]_>*%0K$lu6߳?|v5Z_.]6om]_ɘZz6,8`v{•>|_;Ϟ֪xR?>gԹ]0X8Ȥ.1'_6,cߡ';uhh`Z㮊V^?JEڹuEgqWSoS/)U1Wt ^X%Ns:Z8*K\/Lln Uy1 >'nZ=Ƥ=-Zf7کtBeN{צ4h]MZvxAR/Om+20]Xj*`ћE}¶Wϝ{oiD ?lg~s]w gJ/)iADs @\}$K%gH'c2FY'QW¨a[ʶĢgn4stT<OX M0AZ$3Oȼ%QTƔ IU2C>ozMIcU/rw>ׇ['@?/UԀNOmT* x\l1!qiJ,8D:aTлGP=YF\pm4&\\$Eb\'Y QE){&(\PҐb_}:!9RGt>*T:B`)~Ǔj!^-.):.7TQv71?ޯcquž-&[}5DAZa?i%0T7$fWĖ,Ǔ֝I6-po! 8D{3_F%ciͦmk:R_ܾ|7}oŵu1SWڳ~24Z\[5}^7n36-rӸlO;8r Rڵh%nVڬh>O`o6nXȨQm4xinۦ^gdz|/?'`C$4FZE/1Y_@mx>%mK?\6 ze!5ZeJ5$WPW.Z̓MI4J8Ur?QWp-ij9-:(j.ils-l9浼GGNZV!1(8L9IȳɒB2\ae *d0imL=Nҥ6wx}nyP]GKIl(Z'@"MGK*W;Z"bj7F})ez9M;vKd=[~0pd1SWgH6%_-N QDT z$ VAYFW\t'T;roTZIF%ÙSr:TQC= OaƜ9ҤP`\bd*ZNIkbť U~ޕǦC9,ނdC7ߤ7tt1&<$hN%ITT,̵>9al15{lwֽiCu<.v ~{Q ͝q2}7^ٳhpB%5 &Z]bVFF!G/oo CxkUojc nuGht%f;hk^`I9V(siH[%SC1.VI'A{oGvR=睨XylEpD!d \I"ƃFI%JHJ[ D4~J9eBXE!z \h QG#%R"qfEZ;wԙpMץ]>t[3x0`Q)Iv1%iyۭ݄&?V*- bV]8Yfu?1UMS0Pkuˠ]GwK" ] 5&84}DNЭIOW뛣Rlvd[kkn#7u^r02ƥzC'˟E륧^⺡]c QXBH%b,ds)|dQV4]V`D={ekQbX”Fy&D$RZg6C'e,$QhklJ Rm#Yg^*A[.T8ABN%:RFJgkuN(C#D& 2U S$U?_/ vPx'Yk^&CYefN],.g PJVΆ!g͋E~4\s="T!mh`8ʂL&ʂk_klwgA.4qI)d~qeIGiáCjm]o6Yܲbn:~e͠MP$ edFˉEf,IrBG1xX3?X˲L|+ fTwO6t=B!g~ yAh¸'H?.t2.1 tEɐ<*P3 lUrdʼn^av%0P@-TLi2}n5E1(=QZB)$U9{W$,CahkWcl8;&+H{s龍2%Q;# 4flYL`j3 6 Om+m5DndٕJ1Ad%q ji96{})tYbQ6}K6%M1_^U36̱iIc6rҳho8KI*?%)΂b$eHBҞT$_(H^E2Zt%`ċMZaʁ2l ѡil i+_J2` j #6xCPbg{:cv=Kl/nx<8>^))VJkP̪ &HYed!%B0I) ^lD6 ߼m' Ps[LAU %)]an'fC{[Εj8O[8[hl+jVV{@$J6*Afpf㓬b(Q2/Qi30&TYZ4Ehj.T̐J,:95sHR,GQA,jruo‰JE֦A)tɕhN9K* y5`t Bԩ]^[ [W)&ֻlNr@I++94]k<:)kV`8mH{T}&uDuQ)E 21AјdD,F'ю6JU1"+?Ϯ'iXҴ"霍B^da 91 VJyԂBZO1\$M1Kyh+,H̹xU=1ǽ6.h6- cM{򯵧NX~#G|.դЪ'G%$$.KO }*CH0{VWzKa`xdžB1J;~hV>3^GD"J B|'HH'a68>^wkhmj|^@5ŁqI8LJk}NNU)ɼ_Jv=;LyF{^ч,ie{%zëu YI.rG.+^+ꉏw7Pގ<^n,_j~mp`r7R{~;bn- 䧫j'od=cI 'o7=]lFm8vs[X~fE46gb]ɧjg7}_Ϸ<(W%zMnzVƆ^gՁO+6}5cUFJtgF[,\4v>Y\g,o?_~w_~|? ߽w?:Sd!xvV]֪S/'|~mGC1F{B{kk@oG?]$|ꉺUoQ7[nVMftUs=ɣ gu#-GÃQ9cEM2lcB[ʑOI< _yeQ*7م0KEiͽ/)kvO!.lz/pX#,@ry¹T#K F](djG~>3u2)b"LuUZf2TBgwwa<;.mН2jզ xȤ Ju(4 NQ$-$6{{Ӟc&x >RΈHlH X^l, E(CD^2`%9vAx@!X$ 3fd<Ț Sd+8Ne0LSL;l(Des 4'$P֬I"뒱%jpPC(%-C3(qtT@lYFE %8Z-RHF ")d+} )x0㯙+qpsC&`nVi}JxO"R4.2Ai 7Fռ?{-|Gt? uFgI.Wѽo7/G>eށ1y/1Q׵޿#}80}*՝{,mHx$$y}61l 9cmt1^Y 4@T&I%< xBds)`n!Ǐ[.G.A?DӀly& RkNaV/e<] 61>uƜKF.Һt> 2V] @*, +~Il?^fl#z vq@Gσy8ȯSw1k| /-c\:&>n@F?CwoAZk_^#uw^`SU];_h)j{i45 >d4WLht<ɽ染Ɇ-:ۂOspJǗ֬wuk_~ޞ@[;b; gBZ(=bMǴBY2eMԺ$m|whH$|M]Q&SѰuWGz X+$k䛩eE(kmH#!X$]/è~Y)KRr V)M%JvGUꩮW h"tJTăP%F jkݝqfmڼ&_տ[m>k &n}%>MSOsu sp¹[8w n-sp¹[8w 36J¹[u¹[8w n@X8w H-e¹[8w n-sp¹[8w n-sp¹[8w n-%tCwj܂¹[x n-s|)-s(K-sp¹[8w n٘{9Jp7sp¹[8w n-sp> FZAL-'sp¹[8w nAG :ZRQ8w 2*sp¹[8w ji1n s¹[8w n-sp>GeWj[H>GS6v@T\%((*@%R0 Â=!{L+BP,&efJ&ׂPxH)3:9ǩN>Q&2gๅHO=6 HJG1'8Ǵ҄''Ĺ+ZjB7LksD=S GY yb>\9/pinn]oR6`Z]%~**c7F# ,T U0ﵲ)--11JZD!^1o8ArvxSh^Y6<ŀ?\qnQ7\C1z>Tp5Y"}Ht;c'qb6*sŀ-'S 1rx'b[Ҫ{KerƁURv(D.D+=E̤TkƤWP'в&g==\iSJŽ'_{v>jO1MWU:RJ+&Ze#ŏHyrD=ա+sɊ$s;<>hoVP$bItʜI:ńhr\EaNBP2g5E~{n=&|M{УO#_sSɈIH6}8?<&-$( Ti~Zt h$Lt:xM-(!i_a( 2"8COH<CGAx]k Adj! 4c0KmК3sr=_ e"E o|áz\->]_[>zH_ܯm^MݪyF 7gp]^gnx$j2H2IiNTwHE0Ju1<xzhzyn"'}R*(0ZIAeid[磊 c8l9# prrp5^NAM7ֈ|kE\9!uZl]G!*PԶgs s-6xV*F5OqfS,1$E"vCͬD l Gr277b !Z"A #$N1\8h#ȑ\o|ǓnmJ<ɡ ~:ɡ(T_+!,Lh!,*ýAbWPAQCsM C'^@TG!,x9څܳvK튯d jT;#.7nUk6VڿsZn܈={AmЛQԏy:98nhnq+WгPAP =|Ieg89dj =AՅ>}~%8 vt}K[r37M/ܙ^XH%8k8tc]YoԽLs6(,J5n2o :_1CK WR- :Y护ME9rpv}S ;j.]wSh; 'ZؒK5<:i6 sk\稃tWC&3,vثW!&O;y߀YXhrEc|W_hbrƨ>g\/3C?ϯ/;Af<7Bs1_hC|͋7nn՝(CF7^?׾_ƐR TenԘd6M̤"˖wfU|iw|i6 Fuوw#鹹(}tu(z^ʊyU娲 8J<.`J VNN,=Ж Z g=ȁ@(e!# \CJ^ \EieRKZwZZ 1kM VB&ק 0bT<|gao7|fmQD9LnV&r&XJ\ELT z{;ݦ5u QE.6sVP슧8uoſm9`yL*cq%yтWUŌ*U$Ny-^͓jC{5pJzT7q镅ŚZpIZ[F\|Pb IJ|La%$:j1>14e:?LmĹݶB~ic69w Ac{rVO/<QTakwi^a+-Sh|5fs d3y*\ؕrTtYB9"i K 't2DɐhJ25t`C3a "(БЀ`I:+(-晔.cp &.VOZdn+DHAz @!]"x눬v# Fp\5:j6(7qU| w~K]wUmwE2\ ҂Oʌ4S ʀ3E'=ǦIW璞ǦQSeA[sJ76VDZbfwEAD9Iq@#A;n)8Oi2 }h2Z A rPkrDΣu8KEEa)>)A-Ĺfz2`saM2q}%|C",haݞk`Gk />uv[W=OZ+ʋ⢢]< d&x#K;B F XSHh4wqJZg@V$e&fJjdq6q5qQctRR ,9M9f)K(NHЎ!gI0+mFEOG _d`}$mXI3×dilrz`ؚn])Xe:fMjI2de;&Άv7`f(j)[LA2$|F{nJڢDͬdIZ)!5w&pV_yP8IK"uBye`QDD 5. 8:UjI:?ia<|@#әvTIH)"IJk)E:8YuWӉ%Kޫ(z0-2sqH zVHRB @+K:5:2 Y[zL/z|(WME^j|͋lb3JAΈa_@ .$)R:u2:g 2ք՗CjBPu̮5ɬ~ Qm(1x3˥+RZ"QBQ ]ɞM~ɥg\p͟kţ$#ɦALZsdu(V ]VQD|7HcORg. FVv8 ck2/Py__]ėà"﷧n'gF~|px#F sefkZe`%zJwֈe-^zVqg_FYPFIm,I@n`d}X]v] h !0n$ZDqo")/$Ĩ̀h aI8z/<0hsqϳD} 'Dl1@VGG#jG{Rh6C;tJ뻔慦f;- o/ʂ;//Fmy!xOŤ D6W4iNH{!Y#đbnJ-ccF`Fluv黗ӦKZfREZ|u5,&zK\&hҪs\en_/gOb<^ZE7ݞAoBʈyK#BL^ᷩ[>w\濃vb3MTkhm&^1$$Nd9smvk%Fq:\ƽzkzep@cCٱBrypPݬS FIe%VbO%G.fg78"'bcBd @ iJyQwnK@mm_4z(ыgRQWz?/fm=,q٥ˢ+ K IzEP^Z\r;sJ<gLQ : zI\\k Issa(iC\*Ռh&~ɒ&$,6(h«L˞vOVOu) b)p|S &NL?ҳnAx?UCiGDӰ=Jड़,=܌T'f7i@A)|9>%bGb z lJ9sJd:2fa@g4ʁ4:& \ڝR"%2p#2eeE+/|fDl+H1H=pD}(Y?D8*I@4YQkpgmc#rӒ5+3gG1 _3iCA>{>ޮ8/dاNjrvyoigx DLtAͲܡ-},Y F ^9cl9ѳ8dDɃE@BB |EٕIh3MFk|'H3ȒZOgٞbb fpFClnYH^7@"MArLmt@?+ܗ,駱"CmcsZd2E)1J`% ĴUkYĎ1''WO})4Bg!6zr6msouvrq{y$* +e )RSSzh|Lhb!Cnh;+$TֻcA؜<7FdT9%ނ7`ǣpWxw־:}}iA~EYɳ$ ^GnR9#~&귺qǾy#R.G߬M%z'*_vLvR`)?^ (ul@壶r߼ G=A_]-+>ޗ6= Pb;?܈EM@ϵH<Д"x,6ar%rS;D "]~v 22)xLYJ,po3Q$S [F-L6@U5,GΊ]TVy7cblbty vP/3GB":n-]G ޟ>;A83r[^" Hi%&&#!hO40t,$YwK`l:gwFXR6)A Z'7ڞ9!+ ^ x%D34yMJ0g(IIbVI@^$q7mtJx+(3`a|UӝR%J]* Si6~43>ж#7Wt2{>b7 C]a?Ώ{+?Z~.yq?/|'A POKo@_HL#~O4$sg~/W+b{o,:m͔L?eO Q,p{馄 m<+.O- H0?'Z"s.?{Gn$NKi`9. /1 U^%~=ɒg4SVj.CT*bt6W7^ߵ2 8y?9 U}UuUkNUpɒ#1Y''~l+yY4PM BvP/`wLXLV3.C]?[f[e)b:xٟY-e@E3rW?- g~,!*qm=c~|ýӵG-s VXDjDLRdFQΌY&W5hp<ޱF/D 6 )UhLe5آRz!)d 8PqrBNӫnAg_nŞ9tr7?e^ q}ՐOoJ|xZNK=;Ւ5⃝C -&;%svG­:(NjjAU2HQ($>:] iIQKaѠ 2*IcSJTXdBisSpE&-ƐDc1e]4/٧(Cby55T~>Nbܼ;Z/])ާïd$iZn1f]>^嗕M*=VǎiP{oI; G/;#SV١dPgf=Kj(@Yt ө-A2P@ݡbtBH@ HQI)@{N:\ +]ESD) AvI]=Krȧ DP)!Y6mMLeorPzF /B[{TA5e]p?j/6/^z^~z(ngĂ%pcאm- @J9aUVrY8b؂U e !QF_V&-5DqJs9Gms.f@i傐$U $]ΪXo/jHtc;k&Ξvտ3[Pzd+^$DYd\WAL$,9trk!rV`Lih|6ɛߋ~FPvNIE*TIMq {F j'qx]DD48 Wk%] aC@֥$*`nM4.88A5ӉJ3[m"|?gdRLQ(fTTYeֳN3›4Z[vyN]}z1k"_hl|:}P,Gf,q0WIP92e-2*ec >CI0G;؀.] :iȁaE[)&A/uaqv:2zw oٜ*|ftkI$jɺ a,6yb_Im~fQt]:4u|H*1fGi&֣|6R3(&1N|1w9< $R2 1e4j1dWm<0[b @lm-Lle+%cFJ[.YLƆL= KA6ӓCy1o8__<[v7?ݜ==:x/k?RX׿MUl۹-b$dRB J 0')\q7y(%%EKŮr,!)IRHcL=c;ҎCf }c_G_Vˊ5//y;W|1[l8~M>W"G!IYi}HU-rBJhRRhN l%m۲Ou1jDZ׾Av&9BNP \TDc7~Nx# m:赟ZnEFg.2>Abl]Aze eL6[P:a$ |4!322j5HcHR>;1a(2uo3q}Pq(~7\y;ÎxDQ TZ Txcl[LN)eI>_LJ;?<\=m5r/ʄoXSٗQ%{ՏTy㭶Gԁ se?Bz|JTNcje46ZE5B>L!1nq|ۨԨQIE7m ֠sz!Je!RE-! 196p c^W=cƜO[ T KQ!r\vŃ`IؾS"E+F?j# %4ܠ墖IPw`OvS}-x9oe])v5l-Z+XR66GV1A3E#`Wᬈ@iGQv?JIx Ȏ\I)$ 9*HĚ:`t/_$~.em==X&%Pe`t!zOZ[rH lg`37}xш21zlMR2đN5P|&v2I!LT+/< ]?l[ؓaRVkbL< `Ȇ@Nj`Q \{vyeO3d#bXx$+@Zk|=(>y4cHњ~y )D.Q!Z' Q f0 %"$6&-`4t'7bFZhҋll Lc G,A.b.QY "Z,nŒ&i~9&SG:-d%'r:eAz=$[~s*tNI&,O},fӳ{XL>R;>%9Y>1Gs0P2BI!PGg-8Gq4:Z\<9YV߽OӰfUPT),d>-.(e6?i~gGoT-E_Jzt~>H89}3Tjr{GzJdXDϞ]wl_ӒQ{n )}kHu_V'"\C^s |VwOڧcŻ^g\kIm:/wg aUrYwqy3`6Sp[g23lFmF+u i~(~)֕t>Z?ӓŎ7lUݳYWƆof]?mNlY}zzЗ _eR3HNN{X?aǠuM˓<={e7?y~Տy? ʼ0mA/o_$_ /15[OcjZg^-oXw}@;͛7 _O~2]ww=fQ񴂮8KXl6>Z;y5D/_Hځ]g qVB_ >F'MIȢjjk(/Pd4ŐCr W9ƄWO8]Sl'd)R3d XZJe@ky9[|L|q_ɾΠhw)ݝo֛Po ҝ)qR/KwmͅAHvDFjmRcϷP`(˓(!l r% x/;E-E^!%%$9vE<o:'QLM5Џ{ж8uVRY0.UI+DFC6y/s=#K:G!$aA!p."nm6Ly䛉2 XPYiD![)frDֆA&{YYlK֕dXXiW+h3hXGdV8Z)f+Ԩ5%gBsD'G}ӋCj+حk CK}'똡?]#=3&'1g>p7/ׂ%63Z?{G\@^>/F\cc C^򌠑|9yHzҺ`ؒgzzd&O1A3i$W{ۗzLGt~^X\E1ƢsӶ/%I.BH',tSc.4o>n|;Xʰ޽x[pg/;M7zoj8yiU~_G^:u{~37HXpݹ\p57-9pw2[o;_|/}Ý/}?Vo<|AA_LoqF-3 C]j5Owk>E(s/#+xIsERudnͿ7Ro9;=Y}<-@{ Pׯ^xZ*b ~HA;J>^j_NyK}8E!w -7~E -~*A5F5^]CgeoQd𮼷аj'C&p?tB: B,mEc+Y, YF3r?ʦUUR(0$Д'Zâ,eXLBQ{ كĥa;o(Ȝ(Cp콊E\duKU}[GQcg/uq!lUB/ p1=$6Cr k}uBf!wSW'Om?p0`藠+ Y?K9 7Wg'aﯯ_=ZKݬyu==Fo+JEɱ sܜBD^lfmWO¿yr_*Yq|~Js{+n eny+fz87>g^گFyQo1"m̰dvejŠ39.V96?u^GNѰ>]\\wVkυ[iK"AƊ_xoWfU_[/}Ԗ/-.2.~rq׎Oشf?w#um"NmC{@&BQ@ 92{ߣ[&tyͯM-w웫w?ga0wc/gq95{R xVƃMȢ)"Uٚf}6˨ oy?ܮ~O ]vSZ~}o1O;s)ZբBl T}Kv)D⼶l22 Ib7}I.I_(!9.1k-bm(bJEU?-I͉bcKGNu4JF;t,VαyJkѵ:r^}p֌%)m]3NE5Sa4:I(*)V!Zp- w—n2XɌalJY ol&ۀ)J#I>b+` -Hݽ7 jh,CRm9WIix*)5 "̿Mo0YSh*ciͻCs@;ïB,=z] 3h'Q& |Re]/]kC ^%C(d[5~j`).p9i*K'j/J̱6gD$GY0ڢEA2{g`Mo ~x4 AJNjHH I?ҲSPK\0s;Jc1zּuړ(.zчD v%d)0scԈ%gsiˤ0kJ=V \.%Ɇa*I%CLxe@l? Vi,"VPT@t[ 8gжlguF( BYRTʃl]/2hF'-s~nkcn5iXL :BfC5ehc]VПڊ0i8徠>i0Ii^ȆD_*| >{JC. (JGAE󔱸/bnNc q z)Z wUTIXI2,#<"fЄW&W sۏht RHU(N@ i/5eUE$Ic^ϋu H+h !$ !7wSA0`\AS@d5!PiPY0(3Mr%2Y,ޚrxƺX;C7a*?%QklJ.̐b{e3ŴNAΛ2MZl;Y{S4ݙ@HqHqƂڳ$w"=7/:+ ~J )mFveRsO.5udEw 1R)NUDRhH 2OBDjNc#`V6@۝@VAr`7_&8M2:np]cXvߋ ~)Y䤁:cQBu"wh$%bB &"^gQs;þW+MLJ7#K* ]45Z.U}-ENpE A-)J# I VWe2'Eaz ƒ #"Arѡ^F=Эνx3 nCf6s"Y ԏD?^~z(CVɒ3Lp)icU0 v<΋(jy\E4Χ,}Of_{uEf1#oޣ.)6H/f9 HT`%\Ha("Ft3 7^cl?FYE2:h̊MD0jik 8^ Aze#Kpqџk7m,$fJ$)J!2~\ w`Dd*plL 0$1dD ~i(9eR+xA+Y{Y;5V,e36T 5i[q/ۤf$l@0U}o^˞%0LMiXqnip6a\'}̵D!E#.nnf= kppS:î{{GSp(r2HuTtk̪Fj9GբFCon0& 80[ 3B*3p7LJx^1ȡM!ds \A7fh-:?7 N:)rwzBJ@R9$diOW,X7odhlAT(c6lE^QyRe 4IUU;IYz*=W((c)U2QLYL 3ETXD t أtF$ a^'p `NLUƺӘ1siW 8֬Y T)JK@KUIu2 ha\5gm} ZtZ-}[i>›) @P  {@L!=%% &t@0'N=z%X #*k4@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X (Z}z8x[M{@sewȫes-E 9 7ۧ(Y㲤էѻWg)g|nlge_|\wK8a^{-^7xWOOAaqjL^-׫kfau&1+;4=[Bg5yX?<>δ .RuA' t@H` 6B,9#{e@q u" ah5WSG5%ľLQ{z԰7gdN Z@Z!8CC^c;n)h+_W?zuX _ׅN4*uF*k;0_5` A3% Eoנ)߼v1 AQg!iSětx#XՊnCM"9_6yk.ZLT5lդ\` VQu8d4ٵ~@^5\-](`9B޻k˒?]^_F6)SyҐ"$ǫxw 6''8aĎ[sv RI$k! (Ta]:Ys8tfXc;A8 8%RkގL0Rג2Ôb)PȻ9>yz^AZ3Jfڙq.L&e0X 19 1ZWTBc*1i&?R{ZU`«,X6\;v@] l5[۠oGi̅ێ+߁`Y3s T@oݖ>TC~XbO18_x35AͬLԊΡS[i~g $8EQ.߁" ;=EP*Їd AXׇ4QtkP퐷AmLTQ*<=j9CcrzMh3$=K!Ts8#rS8|R9cA$ ΌIXC͕xZٙ~0_lӷn$&nMǛ7GǝSv,AG>Σ 8JeQɬ2$h\. B0ot`Rӝ^,V\<FJ^+( }xECaVѐLH:2=i*@ r,ұ[f+MhAG|0lXJݔʎt(Vgs/ʃ^^Cyy4IeSӧٮ`ge,vEɾ6iWTkDr;@W_Zr^l"gcVi.ӝS:(u.:^++gJnѐO?bB+s}~cR )wE;\}qeF=.䨲s2߁IF)dg6(ٔ3OEo[~s0!QQO6evAw'yhak1 Clz{vf)FސdD;UZɒa|3I9,wp)]%ΒSzC8eʦ<+ #)Q y/Z)H:K"m]p^%Xt2NZ.e݂ӭyWR5ycWS Ѥ kث7hFI/,Z?A/u?.\EX~YٻZvHAOci kvI!|_}]&\M!B?f| /e)tw3ux:Wdf+Ԫ?v&J+,w7F:-mYf ?(kr(k{#*GCGM#&t:dt82B߇fn0KS'OzM~oX%9\~%Wכ/_SZW: 4@'̰M,-2͂/oc؟~0]7tؓxzQ|&ALҎ;OҙDok\ ·/֥/qX] K}'֓] j EhqYɽc0DʋA|~Gʺ@[v pm֋ -x\ ~6:d(L^Z'\  %qhcwpezG[ܖY1ƥ`v\4X~yyIFJP VTѩÄJZ~,ZxCۧ>m 3mVR& 1A}0A)JĨ6xΨjgo%əSP!%Dv?&X LRKn7eZ46j6]UTQa$pn0eoyryIcuxfjƈZf 1A|f%4ĥܼw*͑TJQ4]C_/{JEBk+$(% )!c.oK#qb)i=yӆRrs_ލe'ևv9n3s_=uۏ,88P#>294E+fRzI :r8ˤ H1<ըUvh5H&PfT4A6*GhNXJ꒠*:dZֺkq+.n-,ޝ-_U&VB'X[iIm&YS,_HԌ50\km$mN1.d;=ϩGr}j'ch?Lgy vVsо,da9`IeԔ3l260f&RBYnP1- wz1|Sz{ijǻv S?ȫ'k/ڢrvq2}>=\d:H},cɞ,q8P*X)dPS;phy `xFՄwKC Q 9lBh0ijc TR[ XH>]M~RRerCEl+(#- ֺ_zIz \ז=%}w7Yu0d^M=YA 6g_ eۊB &dE60AMA#7\͞}x~^xbu󛡿aǺ<zE/_C$Ӣ[sS ZxV}#g-w7f/h+bz_s>>XH](3-y$ޟKby˒?}pqJӛhGXr}=O`-ټYZc}0c'K;lg:~W5kM._L@^W|`睽x^$9E_W);X[~0r =Ad/}[勷3Hx&_6C%1ûo-'U1jߔ HYTGD@'? &Oư!6f8U r+?VFdba4-ׯ;A{;3&N_ns?FxBAn Zx!&Vr˿TM}= L YuңR͖o ';\.s`բԡLPYj7y f[֘vIA&jWmKC3ul0boiRQS'ڰnt֛C7vNp7&W#݃ |1QpD4`/,Z`OFE-Fz7XF0''5D^PKGg- ;&݁*aL92[9*y$qE `;)"2,:׊KlwO~U}ݭƒ54w/_C[t{Ann@wP`q˪S?Mߟ^Qҵu{Rw[nKPҥ*A9 ;&,ύ'W12)VAL_&8NJQx']8L0˦Alj/9vf1|%gS.ށב#0d;p- ˽ۙnκd%ݰv ј/nc?TQٙBBe,YJL&0"~ J=JJ |e3;#<HK48%*@Q8&%R:+fT TGa[98N DM>/}Z}pRt `}ԡF$2Cیa2#ψ:ґsўc>!N} D'$tT`ׁ}٬ES0ڠr5Q;OWAH kp KK5F*)ZHI+PR2$߳h m!]mo9+^|/2|,ega'0d1"K^Nw~֋زX- E&H њhiz[74oFHНkm wm{,/B`%tz^SVsmk_f*nj<9S ;nUMa}(&š`1G(-^C!d!Xr.x YY%r/$wz&,K3bmm(07iAD{ Ʀ4;&09j v\PZ $32Io *BBAڥFxvY3rAU ;@USͤK4NT"šT'w*KGZH〜!S$y"[BO?1[Eɜ_kAlڬ*( 8u8]`{ jNIk /5Ғꂄ!dkt`54.^qQ;UZO<{a|Ҋ> ߏY5Y'5^if_ۼSfGMK[;v1nf=k"XVQmP,[fX1WQPj`\FZ @P:ţxXg"8aE[9:A/uaM8>Z@9_|utc$j*!alD&n_ +ƫ]C,"l%ѴU> F(Gm7NSKQb0Q(..O@ZE>$F`5X壓>$)%xz:cAY2qZS >0)Z NPHB> G,"4ٮ):u&v:kj\uv,kh"}1Rl91Q4l4 k(bÆ#ڽ6rY≯e[T`dmB{뉌 qz1G_=AI颃ʼ[ @H>1 Mѱ#!O} gt]ig ʆGX@DT3LCZ Q=l t2B4! 19[ے@:1i IUD%K,[4]j v*!zy(F2GPFiKV CrUfʳSlD26Zf (!QغRI2&,$@XPE,l(ðf5˝6㓇}@ᡥ[gCq 9V1O7*8wHC7}ykT+ߟ%<-b$$ HdHOE&9E)\Tl[y(9&^|.Z*V=j1%D/ Pvi,Fkfܯav=[qCuo ^>.\ST" I͋<[$R(zCdJdSVR& M:2am>c۶BUUAr&:A m;CQ6Lp%Q |bFȹ_c$_31w⎯^k?[f$J:38s$u镁uMcJ3lAMaa 4lE'M&`O#(>1Eb[6#~}8BU1Fl҈hoqǃEde|9 R@7&lE),#*`gN}̽aۘ2R3uT=sLސ#'AO-iE MR1}[E7WY/p풯GV +yV*[7*5m 6췱=6]jvvtѲ쭽^?ZnbX3(lװ#J,MLN! ;tG&t9y!fo"ݤu/5Mra޻yW}Lcg}g6,ẝݲ]vv]I`VZVIY٨>J!=$z^!( (_C}^Yqf_e& T2TEc*t $-%uDj_F>m @"1MJdmjC2lZ[r [ADniFΖZ~sQʺ7G^#/:DeyRt JdK'amMeBt߮V^ykG:JFXN$^eاeƦAJ` dA m >Ub}CJ+>}}m؟ ú[3YŸZ4l*z =? /'i4}iA29[S/!Heo:frjsX!0i{Aj+H*A:mhҋdl1Ky'>T"v{Rv˴xaB9:+5}S:-O.^ybP^wd7tMfThFI1Uaa_Tü}o7] 3ϥ偼 XwyZ[QmO<&ΫgK{ggc:eBjpZKq>8<<=qViJ Q{j I̚|dq"WtOoS̔;&~P E|Jd_Z7h/OώdvAôUN2MIȢjjBπE,/P-BI8C!9AD;avAŶRLt[M!ĂZ_eldM l3NNnHDwj4;VԔGIݹr26õA" {z wdk n'K ydaw'A1oN-8)kMX!eLe58<a**D˗T,;O1JZ '՞_! kɶ"kdz9$\7'T?v?Bz/ż^6$"UkCA,AQbd"$}nXƉk9sbmetq\@Y5+A)u.)Wn/c;$c3Zdފo\֞-cfu>DQ4lj͗ѦgdST8zFB-A(QdTװ^+ѳ,gŐUT"X;FWwYjLxCJ2D'ȽkkcMpw-L Q@ 9k, ޵Ƒ#_&cE5}ilKZ]< 0T*)mK=}Je`9L-Vc7$J$Ucu*(/PGd6WE2RB\u>xrBп*1o5`vٹͰij̋sKsdBժtQ$b,Dl@1u}| O|!;9RIl8v&8^%$-LT"-/Pyж<9c>'Y ;gwZku0qUN#ƿYfB@d6~`aRJnk&,_3㮙@q-1=mYfK!)j{]!Sd+QtfYZa&쫉n[btUNd{MA#A^2UUHajF4lTK3ٓ=u]3<-6޳޾JwzrЮiiz[F^:vߺx3 zWoWջo<㦛k88r목vW~qy`0{yp~z~ow/y\y?=lw8~:stU+VY;\7}龫H3e%.5߲3Y+] !̭s%ۓ,DקG$ӏL(ȥG ReR9uU|7|8?i{{ th(X9j˚._5qꋣ 8Bg}EFr\w<^urh\{4lڹ_G'0Ygtzu1q^K{ P^S[,(K0  6tH߆8nh>oNMV/H6lm]=]?.g|wӇm>[-SCUE L"js*D;/nTc*=1Z~.NߕuCF9Ѫ2tdCgcTS3'?b 8pf!ס 5'l4F cFnFYN9u$uZ*r"(_`RbD|a]-hQcV%tHن>̺#j7A'rS\H])YDV^){[PjOU#aV-p[-n1!齾&Q VNxY_aЇٜhkgR)"*oKdKO:XNYy! cMA(?M;n%^ OȮ;ҳG#={znuA|}WgO[ZWZţ ܦq^M=_>tt%>{UqjgӃ _ק?_?qu^fw߰}Ӄ[_yE\#h̞ W6ge?͇^L[ ڧ{|6R0b n|Xou{#篝 ɛna}+&;{91Ll^1]ߎ߽&6uOqoPXomد~뒛;g)\-)[ Zrw ۛ^!o k:75ms )B:4tǓ9`g4xϦP?w\׎{@.6hjQW\mU%mCgBQƼ@ rw|-3q:i}s{]̔w { 1 cxz /{mrɠeVQِEF22] $c;jn,LgxV@_(|+aTmߏVlNOr6Q5$dha7[}֑hAhb,No9]֫"KWHd}!r[;۶3Uʤ*,ZMH}4ΖZr%rf[pюiZ8mFF6FYk;)cKdn]۪@9_U딣"cR YU]`tu{IC.{]&d11U)#U'75G :8|M1al}h gf^I#iѼ(%r, N&!+Q)YC%}37BČ6Z/&Yc :`ўE!; x{(\]ry#J|-mF<4ψK˭xBK v ( :Y-5X ԵX lg#{AvQ7p@8D[WaAm!ty$*8ܪIqXL. B2fCLWΖUϘcr\LA\ +Զ+ERZg a5F6&`eh%m9 VQ 4g,]؊~B܂N3 *pWO,82V!MM(Hpy6p%KVGI0(!!4"P.SK7; H&JrET|j1wLՀ\I1#hF0.@Ơ) y6jb(4$B Lh-,MjuR"Ɉ<\*ƨ, 謱#f no9!M6%LlZ%3ż:\6Ud&UdY){sv 4@HqHqƂ,/ڳ,,"=W/: #UGve"R͆.MFE]% A/K&}65"f^,.zўjUa@H/GƒDji"j8(% ȲDDǍ[4k1 8X+pՕb1N x3V;`]֭b_,|[1쑜1ƀ+RvNR!&x?.ô2 /)q].f% _Yo54.Dj#ja&E0^ BP8|J_6#s[ l:2 ܢ0̹(1dȢ N$ZA.;m垗]Cwi`LJ"4oY5 B)U-=X:k,gEn$̂ yx9-KpaҹdRyc-Ɓ{yc7x P>N kVlbi$Ʉ-#=J:{n ۵{)w)jY:Z 5$޵祪)UFCon0&x=·\fҵ5p7LJxv7t>fds\@7"Kk=x"b j=c5X߲YaݢIQ!|◞+lE^QH"Fr.2xdኙ%r4RΦV*P(hQ08QL =AT]XD t أ(7h#J1^G. ؘ!JQtbISlI҈56X~noWzqF1Հ]mj_뽸r$Yo(`SܖuU~yaiX2x]`JfE@zj/$[ѼڂVK:/[8poͩ|Wuw1ˑxvøwUVO7+y}z^c8^OÎ;DgS\i=0 %QpDnL\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uU\uW\9+?J\\ > WU\iW @' ^(:PpE{W9X7w\fYpu;W:Z)\f%WW\).V+8W_9Γ\km#Ƕ"iHټ_ L码~Й0}X,iTI}6JۺYlfUT&\\0+E hZ9+(b/Fpy+"z+ /;\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U\U^C[%6\Ͻ SDoubNۅVsTc57w0#I___@ tY(p{=I'g7fr#_/fX-BbVӡW/=8 [hZ R&Kbgn#lݳi*xRm~ƽ}鼓>u7+vsFCC`'t~5 X#v*wjT\`0)H/ x$F`a~ XTg6'=\: bx&s^M m"ќAfe];~=f8}:l:_7UXbmJ>˘_lm6VgKgť/Fb˥3U }qs߈,Y7YzIRYzw87{HM̲}6՟| `Z?B]V29U4E2 IGD Y&AFYGpdGp6s<h4k<Ź;{\S4xY*Y~.>ŏ|~^ }bkoֽw=ln4~y]nEfIov ,d)zsvdԔ3l2A3Aa $R&o ~s3)=]͉gڐYx0OY{vz_YD\2NOKxL[Tx 2K!NP )CԂAԗn\scTjXi.MʨuQF!$=Nm $e~PY!G͜0XWigxSwۇj9V_(2iV 6Tjx>vL>ԯx$70ۊDB &c!"=0!eQM) dz < y0NAp"l C.Dآ 0^PAv1, OItݸA__:Zgό/~+" bJ rP$eN͂`F)yJ$r*>/>ߨ_,Afޥk#qHdhQ"E=n L#3 pW%DUerє0 yw 4mF1#T;cJ43ƥ4Epds@7bf#o sD rfh&YJ[Ƒm(J އhS쇦9N^ BS~J_Rȍc>|}|๟+軙?xԇ|~L_@z~o'8(2-zeƿ?|aeĤK93hz%ןH{И%=EJ.`!uɛ$佮~:/kvsoy|}}Ii4lwhu3>dsft0߯Ē[skp^5ژfӹKb*?/;; 8.e:Wg!=شv4oKx"Sⱁo%k ) שn0t.c7ur?ptNk?Q7$_ua6ˑ5[ɼZ&+F'V+rP/cZzN%ېU~ühMBwWԕfm:t3k[*PlXo2l$6 +ߟlh=r|ŠmvWL(]D%:06h*w7l.输wcɑNFX9(+#t-7JmxvSs1r1i+rӎrZg0*o1r@Rˈ`e#Bu}d9Edz-k atӖ71gpQ#+W,i&6AH)R,z_:L?~ l}R2q)j o/NXƬ꘵d6$ELJg WڌT[>9Y Ɍ5RV!+$ V×dV!#Of.*;\(YE'qڽqF02J^D$HN<R)!051DK}Vcƌ797|LT~|pz5Q]X>٧)zMKOM,~U j3iK`Yg$$mw%Ch⾫eU'.T &ެNƃj/9vVR1|%gS2{ uC&]ZTھfaNRɴfy:)Lj1Fh Z{1[,} *>6n'颉j[j)Ɇ`n/°7ri"Y;~9 }ݣoͿp`}Uҽ~$:쭁UUC 'J>;ȼ8eeߤiNLR=Zu0Jsl*"g0RHC+KFD§/TEl'ɞz;IdIz*>Fi q*KK^R@DᘔKQ)|P S1~0q29.%mhƐٻ98/, _=Ο0"M{XbG+9d<ijGp(Y\Xi"d,Z ގ;ug]m$mkB<]֛aw'C0!BpoCKrY ZFL&ZVs+%"3mfΎ97B+`i%ȷ| 2`pQ̀v"YQ;RSά~tQwD%]QQq!'0H$*Ű[)rj6sЦ*- $5|K ~~o /]zN5/VCI]~|E'X V qsF$PϺ=+*o_Ro-Q Jhlĩt 6Ub$dEFd>Z0$|&oW12#XP1g&dFllG|\%()W Uf,T ,dazed8K-u[Ԯn@eOAV`tJ[u\"1d w`Q &۬M !9{" pJB= fh/IH6хT;9ێ~<wt)iA7`S/^=q9'D)"raR{J4Ku1X0p㤥Y0r!MZ d QpXG M ~"2 b"ngBxD|3(A^2Or;Lq'9xoVz+GN$p*(ʫT!!Ɓ'YzSĄ3[6%KsIPFbv%`u!",E1uf%"ΌbōI8ZF Ÿ$ x8AJP G9!Ɍ{;mmȭ/51Kf|FD=W)~g^"yxf*ɴB0 ᷃ JXemTye=4-ប@B2vܙﰙe~G"8 Fy 03/ a1V3Fɣ6u=#=U48ZjʈhA b Xy$RD33l9UoGHǗ3`]=lzV~祸7F kJk`MTrN[uW0;Ebq*%CqC[ֱW0Ў Yމ7-w}&U '{]'%J95bJ+#Nx '=&Q&ceA>%SkV+r"5Qwg\F('i58쥥#IujTiO$tBI ZE+T4hipxI`2w%mK  Cƃq^KBaTxXv߷z!)#QТ1`[4{:J$m6.z0`sy8 a1F7i)14X!HPG6YnQʛDo31hLAGSZCra:f9$fL"Lc"![JL8Sѓ4E04'GyIѦc}0B}XvOD3ҨWGvEzYIsE \Tkן6<594ߧo$wg:3攐|E/FRȐ-:RpJ˥,cLaK %V^f=cH 8%I 8"#2kj*…hc/P`C 7ygNt»w^$dBW.ν5N[ }`[W[ww98wfa_P޵z^5~bЍ htGӻMwޅw'z=qtp9cٍú$릻Z:[3zϮęR:ts+' Aʼ _[ iei,uڐ8TSb%q|*)ʘ)BeTk'rR'AJ8zD܂Rx.WT+x.|%,O)[F rG7CsF rSJSA'/1p@Aih9a "}wQk,kl:yY/q a`ZK`9q2V`T#GCFe-brydag ̺ ## @Xg1RI|k*װ|f @g? wi&b"ahB¹kJ" X)ZN{d&NtH2ңnE.A >ʘѤ!A |ԣV E&dYN=Ic0>k>KLIa/ϟߗ8-?s촘 -[aZq1$$ia.͍23L\oNaR9mSC$XM[Rm'@-*-a+'$}&( ѣl! ]޹VSzn2K?ЛLgFn b@Ռt\~8lLE0SN󯑋 W ẲPŕ*ʾ[WD''˫JNDZiS+ eTE9B7iѺ752S{Mlz:82j/0 vk$O|_]rq0|!iɪƑ>e0qh0 2b:p*fI>> =YٿOx9t㨌=jM6+bCcNjK$F㐻MJE-Fs 9Vu*S4ۼp $ǣ_W߿y?_(3g/o^9B@kSO¿4Ο0[C׫0nkhЂ֋e\+7{}ZQzRV ia?Q#;[VR~]"_i#_VrtEլ~5+mWXbC<4E׾^#kF*;uhAzkE.p=Is-hb` 8b%f"&9㎵)DA!$'J=#38|tC}`m r3PF/8u8ǭ՚L0Vίӊ_2޴FIxw> 6^f Нg ;OR?n`ǔRkJ.6R!H`"'Qs"yHyr&bPL6aKgUpaNً@<m ݺ$$+`]TN[ϩM#%)_DҀA E#DF[}MO[\ 41|I"Um5p)@PUS]&$tPNo-e{7(Jx?C1п,2G(?r4דIQϯI]((R^~7K?8*&柇pTrX룍Y+839:*N=VwF~XOQTx(/v1y^+EÏ鶐^vGY^ vLHEsE_3,d!s>=-~ţ~V8şM!zN8dfb{U/7(^eo!.m&!p4,i6(*/\k֟Թ1|zۗAr6Ԯ3NK=-*@pozLoh[=֜AۜqB1drsr?8/ţo&s>vkN܅K#_]\>ܚn6l,W]aBh><*otdęG3c<:; ;쌎9}wmA[KvJE{./"ٴ?}_F?zyX0W熋n-}w>*H< ;8i>Z2ۏ2HǓl!ʫ?,B\"; xf<0_9<LQt|։޹-cAGdѮ9HwGF;$/d+[O~v=9{he%Vר3XoFz-/ `[VHK숴eF+>GZAMjt8=o bˌ up_8$O &`R*(TB*]YiyTk#bdz]9fG4֚]${Xzr͝7ᅮcp-izݛׯ~~1y69'҂ˋ|c$vQOdZrȃӞgMTGKJL&i|7 z.{9xz$o?`8xp]p]/i/nvx!-?v~İ/߹'//abq. sf]8r:$#^H(p?S;H-wݥ~$2ΛF?7P[59i-,r)ḓs^fLUb|V8u.L$ꔹSVuʪNY):eUi5k0YVu 2*:eUUSVu ҺeҢpaU0 F(\ pa.gV…Q0 F(\ pa.…Q0 Fx! G XK~fB<.~Οu )Iebz =+7ژO'f(- 䜓{T^ʫ_(Ɖ+CD^9ΠR0@'E<>`PWꐒc", ͙yeEI`Gs@P}(} }(8OO-/VQ6c#rgFCOr/IMƋzNs"s%Z2hN,!a%T%P4˔No+۠`ܴ^f8^E$gP{ɮk |[Go4蓻I֞q=8D/O[j U3W8Vw'ljDTc:1ՀE2"$R3u2RCtR&ǥWt1Z3ᒴ6 q.s 1$˻zmT!,SBSG1'ƿL+Mxj&v`t BҷCHk5s+e ߧO.c$HW<1V: qBЁ zIUJ}]4zA+B%%BF.I-ro9ҶKQ 4gZ,&efJ&ׂPxi&Mq D]]{z6+LxKt$Wэ$ ^h59qk!'tm|8#kOSbIכܑ9 !VW*阪/GF~ldi~jüZXT(ŒTjq!^1o8ADxj~~G]bOG{캱)4glz3*8N,>$|t+"g6*KM[rVh$`;C>pKZ`oL8*ZJѯh6H]2j͘UZn+q3`0Ik"ȫ{O־Jik꘢(7%|Ȓk)K`ƑQfy{]Xo|˓%yR}~ܬt9{'Hi;:Y.a?-S( 1I@N⥂>zPz{fۜ9-:yqt&|6$bZ<½x7c(2"8COH<CGs:w)23ʄȲ P}\=~g.|V~cDtS|5ļw`4Hw:Iury{t!4` P&)qN%|TN>p ^t;l FP\ZSk۳ѯ6b? m@Qۑ]m ![Km8lM*F5qݸ,Yє !V3'"Ƒ!$8kE$$Env7t3Y%u IFn=FqpeVxBH{%TRR%\w*M_;8o{fޙ&MSQ\!&8MN:og|s/_\t,:rQgI#FjdFG{8ڝ=EN$d0}opɒjM&bS,'B[9Eq3RvISc7̈LP0C8v?tǓ)jIu\J p`*rBJHʬNp)s.EwTb^p)&)I2 *YVHˁ1.xhj72+uJo=5ѪD2uAybPV2!nkߪD-n'gT[hIۛcWq\!|i t! [l1 HxhڧGCPzoɉ+V:ϼ  =@?N^7u" [Pn)X-ҷө5˭%6ivX@BJs@hЎ[J47>f&V%N{eFH%TrDΣu8KeL(,IiNE>mnَ(89&^\H#a]_2KNԼ|HίZ ܾ#j@|k{ճ5z\]Qƛ $3 [G0p KK!hɄ)P Xnkw_h .vvh \|PIЌ,j\e0FG'(¢Ω 8SPTHЎ)MX~%\yϒ4dbġZ X޲&ΎvV՗g`f w?-/yrBӑHFdOd2*SpBQ11eڅ \<o~_(9w%gAI`9*DLSI!R ,SVkPOz%N*qx;%-hF #Qzgw r*-bKs F8q{֚N*Mf*ڼ;AxaGQ8U}=kMEnk|eb12L&_ s\֢ؔ @DHI!F/U$#h<`}4!I!?q|VfrEXw}|tmI $r,!$N/hoL8L ,X'r^mE(,@IﳥfT9K4]|2J (yG4: hlSP\R.8rYJ29qI `JHaXl 7}J&whH b x y˕ג)$PByo57$Ly!B: QnVb:J[1 ~%عm5VUx%aOڄQh#&9uJGgs @OFèp/&T)C'hU^1'L݆IcA$yf 5LғCa w? -smM0JI(i X̞gRC rDӒ94!3IFꖑvla*>2ť!TIˢE5@rZf$)e'JYb[eiUQX*? _ZwgU? M ěj6,Fb 8i5zפoŒc$IR}EiTar#~wRlo+ͥEj޼ccʫ⋗7{&VT{y?;ta !R2+|x; غY;tnv)ankbĄQ#h*|4\ =YٿO7ܜvT6Q+.G?r ɯ\K1j4ܒcM U_9ק?O?N)<}uhic譛0<} ^CCWCxҪ-V˻^,6.oy͸wG7{k@R~a~Ob]ף8[V&~Q*?i_%'U\twfn%ejD%b7E`]o~r_U s#FP(՛,3I A,JuJgaw6%s9&eHwѐJ ĒmZ[۞'d!QV۠2z=Kk*j[c:["hQN'~ʢM|%+w|)o2j!tIBw4}dqBwVf6xxa%jkjRlmBN֖p.0Qyޟ`,yj,PddYzbeU(?9j|?meSX;/e0]å?LGW|[MFiD7Ҩ *^1|}uǣonuݢ]u2MXpj4ޢJex_u޵k]GIz;^Qn*Uc&'f4bXܼTR^oqRRDeܧw&$k'MI3F\^ˇ(=w"/o!)\`'јThXlP5_?ԥTr|sfZ9ZW)=N˵{S5 Aх0xLzL}RsN9IB\\k_ 2ѠZ8>&qg1{'4Lm=Sr7<&͌-?;GLrxז6Tٴ!Bwj4YҝMJf `囤Zӕ XӁnV]z[\;[k'oC? eC:Wk k'BՖ.I9`ro)*p,BzRc d%OI{+v/~wCߧ<9EK$Va"ےH. r :4x=Xa=:mDd^ cqrf,hB>;g8xd9˝<2,˒<8P`)M<9,$M$Z'N2 s=1#GlWk>{~D^4}~/daБ uEQ"P[P $-2K8 'lǹ< ʠK"hRvYM|I:\|fr,PpR2ѝic)M- Pe==p ( e ![R2uS<:ޢR, gc NZ ̉o5dQGH41w6eib?% MjPRr:2/IEh?8Jvn%mrP YYpçD̫d0Kce"!qR },s Et!OԿnԿw? ׅ/Ⱦ%fhrTn0 sOdSF3ٹan^GxQN#tQ3!& Մ]$Z al΃B5 ~L|L}118k&Ƃ\zYr :Iān6&tnJl4: l@A+EZUfR9q HӀ F`MgKz/Pk]3 XgI޾fz ouwV8(^`K+_]^/GӠ1Bkji][os8yel:5lnÝ)AfoLhC&=oCaΒEt<^s.5ʨP5+rږ ie4{mHŬ6P;dmRٻ6d (AR]^C֛}^ 9CZ"LZQc,NJ4Պ)lQ3鞞:~ut*fĚ|yG<,X!#.!?O,m be=* Q*vqG9b9zDl[占bŸf-eSJ^Jl2  ɓwt5X_nvxzruKyΩ>{u?L& JyWI)Ltnέۿ}ޕs]!o~jΛCuxYZ3W)璤' cRfzu~vgOn[ۓ~\3/|A|k(YԂ j%3IhoBgǶEl`K&E/W g D J+dB&FFPʫ6%FWjIuc>k65l P֔2ND:- H%QG+,^gL Iܔ;x \.אӟBGɔo*[u>Z.Q'L(afAt>I;;dwgT" -":Ґ₄0g(-4|& 6c7ϤjpU% Y! PI4X㵗D 1Me LAq[;ϭvҠB3kF"2o| X̂tZ1SQ/5F6` JȞo( wю=`l ьB6#rgq/U0Zɗ KyAHĮEB.뾤.dI]JYpRfDb`9p*P,8A,Y ʃv(ϡuiQ6"X-8Df+(|R1,aL4 2Nk R0ւv 8hjܢs-Ry(b&B(7ͦ2rN;H<.KD.CٱE ֍ɱ6(IrB<_cv ؄eي&[T+e;PDFUO\"AFVp<N[n4 8HPtDQmJ~WK,~ay-~[TE.(&W{{Dea@eIѱDȵ !]"{jbmwa5g|va?(F@AFiIF }gʳSsqFiOT1g#'T "3 *ms m49" 3j <:wưnVp˯ld[Z~mAvh>rFq1Vd ²='T{6:jdK%FEKŢK`.>(R e+R}Skl:KvX,lFBX,|RYx.EE+g',40Mş]0`\bD>*kKe$ I:2a4m>xǶlkUŦ~Qm;|Q,6ےt%v<,d[`[QǦR7~'w>Tn $J:38IV1DJS>d*K laMa1CI` &dYI I%e1Eb0m6ao׶gLU0nDlFIDB4]o'wqţ "#d.՜)0 [osJY-Ktu(d堭M%wR1RTr&%+QK-ifEPBR1nl:k4Y9&~ՑtQx|ͨdS(E;xmuJg&yA:0 lv5VuіN.>\˹ul*Ucya'a %<0gU0$]'~4)z|<$ͯ>i3Z gʼf P55/edK|DH0zƧMp.!Rc 1bal3Nl2u"XTR6'ƶ5YHDbxvx`^u~7s1a,3Aqخa>APk{]`sr*;5mCm[I6R΀TDY )/}VI+kbrҪ!Ll:3u.5Mr>fqS=PYdn"5}֬Ew㿕ue֕z-eg~" v]p`VWIY٨>J2!L2 yPBaHqW9Ga%[d> L1G&d0"Y%FU'k$-%uDt/#¶ElL8{(jtt>[E LxdV%{[lMgMr)M?(ϲn`kvpJ:(P3ZTB 6al$6lUQj,V ]y+ (R+aRF:ϫߧy0;#(0#զr}*܊},?ń{D:d|)!wD |΀dd2"1t`1Pq+'wIȿ6cݩ?K5H?'V!H%0DED[zY 4$c̤-#mT +{+A! ڧMh҉B7`\]NQƬ/̕B('4^%u?[z{jEmd/궫aqO5=-MVTz;qs3etl TPRH-?Bi?LtG,Ois|^@$E`9d{Rg 0 r1=ToCwzA2_%^ٗ%ы'W drAU<M ]vF`]q\wD^Brtמi~u7Mw;& /$1x_ Nj*Gx#!Vʑ=ɪaa+J,?3#fOz+XWjxa1'{N+\9xUzVƆ-GV?-xl|19_LY~z& 9HM-į dz]39χ߿?w?y}?J_ϯc+pN)() h _CcyCϪܵz3m+ƽ>^RzoR Nj?F>,cG \jMtUq=]^|l~[5ґg[B)IEc X}B7k#]Ge'F$ڵpe E,+P-}I,! >Yb9tBmaޅm d)3.BJAP#͚l$j`b"4lUf4t $kpxejIRwӤ\Y+7J ʢ"UZ0lpKv]2OĐ'Mf{y6̣PmQ[ H ۄZl%QkEQjj z/s)& 5X+e)J&63Fege%fiEkC,,23YEbx%Qb+̶kZB]]+NtuA n@\RU u{Y;t㥋N0.u7=w/;|߼k'tDtyL{+o+wzƽ_{do{8FPd4L(S`rړpWx}?!jxӑn+˻\~Qan&2^3,d{_~9nq;iYğO/9C'ZLK3&m7}K\Gߏ{u>ن-dn69c8Ggb߅LP׋ __e2`"z{4Jt= X oFLoуwFL{s&}p02 N-~A+|"K!Y˕T ;G9qp%h''Ş_|- kd+óyXpg&jx,-}=T?vxO gRbZ/ב jXhuo:Z-o:2.LꫮwdŐ\{:Ǔ9lPҡ?K NH28dNkt`ÿ%AFIQ,u;b| })F5v@Qjm#@S+ vFL೰Nue4YȤÁ;03&IR Dº>$"Uka xE9%dӟ9ˍj<5XQt-G]ksc7r+|VURىkf*6Δ @3IUӸHI( ;OAw"k00p:Xz ysr r$͙#y+o gfg()3qΥ)ڴuO7=tzTEN'"H V%Cx I2(2%\7'M%w9`8+.jPrfF-iAA;8C5Kuz6X[>rBNZB8$n^|x{"![^@]N ֲ6II"zJY :TʭB,XU4mZs+ k0BeNl9s\CQHOl4, !e%w(KJ CX^1y :e h\A[쒉ג+hfHApAՁ}F'eB9|rhAwe341*g:39t>}GM§?u  h;" ksRh3~;g`bt'Tm3KrDBKpӤq$f. ۠}unݔI-rL@ BpLJDU"020A!0׶5gG zt(ÓwOfxOoo;nwݫ [6wK R;;}1wk n_JypwT,sG~>MX`Eݫ67j*|wEKC+%h.>Gü'e)CmwT >ceZxwKsE]HJ"Y/y%]^gIXnmLxĢ(!e\$Jr]8o-*$hda֥;te+$;3}ۧP1Fʊ "F(Q_$f6fjbWOK\`/e,R:ee_hjA@1[ h 2`CeZsO+C2RnAo ęVYnmZޟjp# ss3#<^ J8M3*˔!iHIɶ3D.#L>rpt=LZL{Űٰpօgɋzߌf1GD= 4@o`56`ˑWht|H{+;{ 5ԆWq-4mn8z?(m'޾nAWēQR)|5ܸoB"օQ`RבvZ|@[+ູvHyV+Ŭ?g@xfԛ!ч$G?, &a1euf4Lqt4H $xpq9J)ӧ/c;tƊL^UJK[%Mƃ|~}_n6usPl ?9Mcm+r '+[FV ]7'퓟0v],b%U_TN*Hb2#AG2Wjfwf7$;0 푀8qFI^%ǥW|roΘG;fvQxcr\X/'SK/ncʕ[*WFkK5$ e:gd`L+5"C玩㎹Os{gG w} ,CöaA JYCY6<˙<=qb"P8ْR&`=JT{o3XmуEi}|$hw~_4|#/~\%s4\h-EHeʂ1qBrw{ɓ"1 TGj%ot5a`ZΜZ'? ~>jaƣj AH'%8x*2N) @I5vM—󁯠_"T;_h;Le8wMp1HB3GU&{L 0O҆8JJfm΀B (wh/*.؋U^ԿRMbMlJ2/m6\JƠ~/ƎWϷ[ TM#T6҅HAV@g@#@ +LB7D=㊱R:w(59jĘNQ< :ճQtk<*yzΡdeN@3>ymu࣌ w)pAYX 8Ysx npъrQ{\IKn ='7J 9yדL:r@B: u/PT۲뇻{#+Jp[f xv4skVgKA MD)g 䨦;ESV١BHj~g,1'/ ~ 5?03\L vhЏ{WW8o>,Ǩ_|Px6/CbzYCF1>f/?|ٔ&2xT^M0CAh;>pѣ ᄛ޿Q׋8ӆw՜e͆+oCI .S\7|=0g)f|Te_$Ǘg^-x(%p #8,Xe,*UG>ό/v^?&oʏW_t2G?OFQ N~}Qyv(U&ZAgk}9zԼGc<^~ 7=;)3v2qu4Nn 6QmhBZKmq+$jlGoFwd3."ME{.bf.Mؑ /NFb\$;%N^bCiE[ !}?tL Jս#zCO6n ImzEN S,+251/6uL1㎉VU{u5<v;goxthۏє7QBg,7vOZ{=F#W>He物Ps}̡Jg(߀9))¨CTVP iPJG:K;WE)PW[J~ї=n~E\xꐓwJM3ݘ6[U1({Tl9(!ѓ{545"xj#}ؠ) cbBhO,<_=JJ ً$glZx2'TҌ:,KPrPfm Tttf ^*ZNq@\Vk.F/i\ƃg݉xM4 :~j`9}λ5@Z<*3UbϐpcG]w`G?_KiQBGL];C_[g]눐 Ka+ @*ŬcI9]r9gRVR&jKd=ьrSAG@)FqݾȝZ>|ឝ_ 4urJd:2"(`L F9)d%.k]w[|u"xZ8[*+fU'w2@)"TXeS(45:ᥢ+,efJf:MB} ;mo`r( EM+w9k(vҹv37jif,dzzp\҄PUjQn @Ќij^UcڪOfK8pB""/Q2>= TR9c Y9#T3|'Y˹DXsژyrSZsF׵!!$0K[;ޒb}4rx:s&Ag>J}`8S:Q8H$??*@ jAndT%ѪNtL[͂#;;A;EԔ4)k *;| d(>& XTVjޡ*(54YЇD|ĕʶ`{S-9̧ce`eL\WTuAgFcg |d~tsbe$JJ3$3Ss L2NH,T|:σh?ؖXm֡nbR|)#A(G+IM>WyduR} ,¶Fۚ[bw04B#1;k5>{~t$ K P tpj!G hNU%8:r;cW0ޭ"(]fz9Jnoggܢ]PM'|may-k6;gfp~Ëh7Zmκӵd4Q!t/'IX8*FxHgjXjg:Jƒ`NBX%BL-I(-$[ƿ-ʨ6[DzzP|Wo>WasmzU^hc>XqM\jc le>,PrRNVqWve"Smbh\_eny"=9sIRvIaE#]낓-ȘTY :xQ{[]]- Oj[euH/hjElbˊR !#ZZD IhoRuqW:DI)Pe֑#J"Rb4!\%mo}5{n*CƋ&E#iU/B+3cDeqlb~k_UrO?dC/oQU( J3?b14&sIN\2yr-U"!w.?A{\1cf#ԤFdLHIvLs- jqRTZY/$2IE (_6!6%U#bUm#izΚ|lhq֮ƟaLD:T%$ Q:Zᴐ`2 SB=YBsAoH (FNn|nuzzz%o|Q{wYDhE`(;FĤFh$f9! R G1;eM<}3ژ u/cta|W̠ 6cvr>|3bxBlm )|* em,lQh *19I* B%E&C#!+~C H)m'9 cd fp {Κ۝"+Oœ7fOCc r/x[ڡ/x[`MYj!ػ(R6*'YP)[L-賮iL&fFݫ<,b> l!32Vt҄,k2b\ YFs1Y$6 CyM~yU0nD>6D|@DV$" 2BVRY+а1GղDPB6\~k{'l-}(5N*gbY5ؒfV%ZmgtN:\s&_goT\=E;xmuJȄ轨 m^ed[!$]M@{]/)!r:6gyaGar+ >\NӑkL0SnѥQzqܟ~Lُ/>~t]]ִG}\:.DЬwO,ɣ 4ޓGoG?;YSd ٛA2sѲ"_*[T o1 8 l2 CqM9(( SفhB2D`4 HEdkp}Vk.kbݼrҪ!K5,9 r5Lx`%_,+޶pI7݌sV,ݿݲ3G>lu+ҼϞ$RЉm) 5" ]yP^l2+0 *dɖ}T מ(: Tyվ| hc1CDP#. .jmfR#+<,ًwīiX.WWXIX.X|Z Bpcn*Y !Dt6L(k+tߦ`V<5ZKc#` +aRF:Xtܠ8&yR"sfT@[[1OǎpiDY|yuȟ ,:cdD,`#JcIױ+c"WNuSMK5H?'V!H%0DED;4]j}5cc:eBj1ZNa:8ܟ<ݧo|bZA`Rn2HGwtt|TF] wqȔЌ]:xCY-}k~N\?j [Bixo>.mB`]qR7LTG9i~u3Mٙ݉.O_gL^Qix?dnz$Lj_/ ?? j$oWNW V\9ubyŌi4d&=sp:8•*q\5gelriuӂG/I*M}ٌg1XwS#7Ll0nq:z/~:_ՑT=?z':{)b_5e<hc%Ecb&^*FgQ%Bl&,$ѕAmiSw0=iZzj`pFU)4mjޝmjE_OM h1.47 !7ے|>ن-dn10DG񄟱\ I> 6.aǕ$UIT&iM II]8HVE9NlQuCLQԲ|~y͆V/s?} 7ǖvLA;޽1S7mNRfsRRoܜ[$MƠ'U+hRw.ae6'řGWv.P5Amfʎm}p )$$G֛\LS( )hK3B)T5 w"qeO }zcgf|471sօ(\Xr NzC֤xEjd O/gUǀ=M /Fs!yacY@l"yMKev^JT*1d'M'c'$cBom}+q M1wǜZHѺcn<I5mLM|@ d^i4Ji\(H]΄s&(y98-?T\lK9d%x!bR(Syɳ֗" uM1bBmM8.qԁ @ /UN~(I` d'alk9{Z`Qbj=?Y s~8(yRӘ}zstYVzwg95_׋%42/W™+qqQ9'Dg.PIb>Jbq&1(l -ۧ5iB:%#m^8s9-wVpq8Nwץ5NŒg] [2f.!yi)vQHS^cM >Km͚OAgg7KrL0VTP&LI,ͻȦwz`G 1 .V1PԛMrď6S3]4Ng}hV*jBnѡ1c2[-3:$5pS!*u> "$l<A*Ї:גdTi:Zs@m3|2b6-J\ۏ {Ѧ^pO{nغ`futxHxULim@6mjC`Ef@ļsuPw` WRR*ZHGZO)̳te,-Ki4J R_C9E:9n%OAWqn^4ع 6*ّWSLk!Y5);[u~ E\x' *me '`A`-;Zk/҉m>AnĢܲ1="o X*Q{z=*|>1nwS~}~GKrVfe%+ "ԭDNa0i.٘9bk$w1)[.&{!H]:C_B(](RGD-KpkyZ7⬯@kִ-ޫzbM1ݮ,XNn[Jz%éyG"۰# E,P'22b LʎWzOjG>Qm⃀ ~ ,}Ơ},*{'i[ \4&(qdV͇E+ _egQn=pX!>[7}-p+.|< $ؐ@՟(#E!) 5BN#D#Թ0f:@yHk08b1强O7Qb4fT\~2sr0%_%\-?7ܻ6t8}7 +YzU,yGS.x86<=5Ө}HTOQŶjkeGU'xUa5ʔzSE9mIR{sl2ʑѻIӿKM|rûo>7Ui7q8ENиGu8@c Xjs7C}PbtzsqOb:"qW[b[9j\k>cB{xwqޙ KtbK̟_@ jAgog+nJ)qe5,V-}yj;60zCD\^sѻKk:/6._қ廋|xw~XlJŋUWl׹K}Mų11xo ~A4fteM->6Pp73鐉S{{Ԯ׷|}eGK>1F?1~L˗J.-߷49%<%<`Od O")ڲ26\Ѫ! O䓌#{5nt.e4TUZWCJ~S"X_ ̸ӬwYZCjNsvls/DծEIKc\>`7~::a]̔ ,{|Tn??)ג: 8bQh[v@#oc UW-n;k?nX{_mY6|t?.ͥȺ*-QMAĻY9K%H79[ ɔ\HnUK%htZBn3θKi6kx4cTgkg,B=Yk)1ˑֆsPDzo%[xw4Qh) mGylNP]mJa?NVP}*P3u-PRJk]$\'{?ք&3q5sƖjh4U9=QKʇ<h gJy&TR4vFו%cM;F20T0|6>MDŽc,c1m 6)xٻ0F>=bT: & a%A{gObjSEKG: X//dc)hOBr>7'M<i9ՑJj MQ F%k &猇s9kJKgBrs|jH)F"P,KJᨮ֍KA_ZhI54I{mg͋j#r}HF 9yR 37gYB ^|n*Ht=) ;*x{%8Ҋ 6*˂whBJv'BX EA)1أԡt@<+h-aK>83 AL7_u]/2hE ǖژ[Ck2q-Ygеo <<ƺ2?K+w ˶ Q .trҼ> [(K9ԠMX m {Ze3o:@G6^1ot56bhqU +0똆'z$;uj,JPV24D(2!Y!XPptiX yt/1# u6B[!q2p /e,0PFu5૘wNu2qnQMZ uY'QO6$dzC8[."O֮ TG<)} ]{ #A|uC.MAy1k"! 5y]%B*q`(#FQpu $$7(ڽX1z, mM!gT]lJvlp 䁈:t R(VWC+PSܬe$+oO"/D(8 td+.x݆icm \g 12iU/_Sd@!jPk;3 5\oFa,*ga1=TV( Aw ?+B Q|YBj^!gѝ5MVU#YP›pPJЎo5JB42߬n$l& \Gc%ZY S|cȵ6/~+]Wf[6_\ϕIv̇4x@^#Yhl)m\c Fݵ@SYf1h\ k1;yHY#h46`료? 2#lRӑpPaRDHZN!F EtNJdt J`]PLMY! A)FaC7obO<Ͼ:D?n?`+AZ964NB0Cʟu7 /oV1Z8R.|.* #-Nu#Mg U4~ Xڨbj#cR=ڂ ڊ[M+5Pc͚ Rת%_ÝAg2vT VWbµ-9V#ϻG&&Ks3xs9jawU\@v*f`)o"ae k̀|12 =>w}!Z) "Q%q\곹R-bU^.M/Z*cQaBM2Uq*btBeN V3f LBF \g'b!(vW1 ՠL>B-▓=eDeK șͮ?**G4[jҒtFXXQXirټ9`}i nfvP{UMK"RJ^FRH 3ds'JH[C/e{g˃)-x:"9;Z X;^ ,{1BVy $($@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$L!zI$C^ sM|1$uϟ! 5@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B @ȐJ$GrHdڿO#` ϞNHBaHH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HHhwϋQSnpoݵn䏵w/ )]ߌ9E^- \7 ҳ>úZ?/?ooY_[(J_A Ǖ=|Xvxz}֕Mɑ ӛh:eM-j\NKp=ҽޕ5#n`>13v=mGN+d(m&K!RDe;$ DK 3a@N~̚Q q%4+nww@ "57G"PS+\zE<HxBchQQ?hf,s>&(As0"p@F l.QQ #SW2Ws1; $w[#!-cSsu{f=TeD(ybWg$Q//f^ -߯cRŤuj[@;3k0V̾@$l"$^o%{)[Ug^7z2/h28Qo'G.3*nS ?H$$X_l7}x.ἧNڇW-OQuӛt;oƶ}8S:>aJL-᳭Y[m-:0MJC|B* as1LVsz ѻ# zTIXxnky*'$5`WCd7ګzwwUnۼW0ߣ,4hrZ$A|fPPcVrX8 m9i\Yq0eS㫝_#>wuW,"ϸgI@萒,EӨa$=W]l!jNIE`f^e,Y4fܜ6_Dsp2-|yT8^sz?𻏜g!(DUXß;U:kv-;}4 ٴ{ȣ|Tb&F&c a,XSRv>Υ' *Vy(o-.X:MoP:]/EX ܓ//~mC^m᷇9£? @  GofQB-`%,F"5Cbu?qZ5/ªOm'jDJ$F'ۛXNgp11JQ0&%F%R$FŚFcbH0K[;*VE*n:$h?'t$h`K}HM$.N} cGa?gIIϓTY(H88< *t-YCuގ.ι?ٗ1z9/C*>r@OW S=W[ʳ@%9Y0U CcZ**EʾAX;HHrG- , B KA eU [S->HNf(C__*_2ԴF[=G Z4)'t.h)3QLCewӬ7/n]oSpN=1 %"~JS1s-Zϣ c'+Gg"`|Dl Hɂ !Z%g.ڜ6p Ǔnhg#!c1 !/Sl) cq6lPF+u&"Ot~=3#NjzMFʧGLGŸ9JAӵ<|=6=zηZ_N뗒J^YW/(A2ըHQiˮv7 ߣR  [@_7M.0^Xeo`gt;`qvFK7UpN:?\gslf;Ɨp5S6@0~,ͶkN $eĝ'$ڹS&B^kSo|b7p4_&Of\4N )|[:l6wy5{pKN.SEN@S; :W-&,Mq?vMu 6ӳ{:NWM5hύ|C#/?Ğ}St::&e?$w +rJc#T܌Euq8' :~fspҎ~/_OBB Pj=?ogλ7o|˼˯p5Qג;J~Wg^H.8b @Z(t)D)tx賲Dٰ}=y?k%v8Tg~l~YOxGx8a'ޙX1/S_|='hS} e4+}/'=;ˣSM2x~G/ٝ8Y&W:mw.BO2_Kɷ";zKޱDo"77Y1En/7&_[h㎻%y!͚iIMtz6AV]vҢ$]*v(6vԹZz:YAMGn P荷 PSjx=wTe1j*&&v~I5.S\~\fTm?˷!{9{^ն%'?Z\yJVEi3Ui~8d%w%) dkMVdG]͌uKƔi_a4;퐜/Zc')A {Y˴&c6-$1PW kLX;sƠSV dm)ŚqrR$P,ԅPXL'&Wo/ެf.6W[}|s<@$Y8bco1;IHƁUŜ*RIIbk0CU9 6@ %ElY0 30sT^r'yPl>$ L>&.ŚpḶaǷ0,"HI g? ٤TB&a\@$3N9V>BfHx#gQքT{ 0 .X؃Q-}{ڨex,XL>ByBxxDTx\@ XoN')$b>XAp:X&. RqG$єqQŢ,iPEfH^:8Tsv#M7T;j:IɡH "iqŭc+T:cpvh38 ޵#q83ko}ȇe|1 >[XVa-%Y[bIJMVX"A@xFBqޔYIn]ź7s[ܱz(Z^L%d|t78}dFѸ&UޏKfO%kRGH0B\ZdkZTuJ{ItqTjQmEnQ=JGe3l\XKh XU2404 5uv^DvƶOilWzG.K=H6)|+G2|Ū}&x4XOv ɱgRDŽ*En7)H$RiiO]֊QŨ'+*ЪDsJmyٮdZkv% \HmB:VC xx"iEr,bDԾXl[*r8$wI` F&#!YQdtY !2tr!? 6V)1_m`Dc̑+P9NO< ض#OklCk>Ѽ'G_z(_G\,SubQ[ZztȏdW|p=Wm.Tl7 +-n,ܵW1m[ڹ%E6:) p =pST2*QBz1h9IdLLδ:LRrȅ`IE.CRIHKm'i Cg*}* \R.H J6eum!*ɶVcI'ƣR ,sδFܔcA3ym!DG-@0 b5 6{Nz"Q/*1gB{%c˘ʰ,K6RIl'g$DK@pB$SANHƮɆ%(5d ?[fVthsfRm8ٮ>Zu31{EM8֊4Z`@% ӳi@bн- e%} vY?}t湜1;#Y2N2yl5X[> MF%͢/ ]Az12yz&r༓!/s[M`uhj~pk|t[),IhiBedYȼ&"}rRE$K䠘NXv!,n=L%Y h0 ;2Tf$%)DB*E _;nj2uqd^fh*ٞ6ic;&ϥ8p4)?4mdM5VIk0Sh,Rwg9#1Ÿ%-g'Ă\5xY :X[f- & ܖht /ԁ @ Ǖ"TA K!'s ;;B%)Nˀ5r6&uzھTh-aYҍO#]O]WkNZ҆D`7 uu=n{Ngmv6eM-u>7v>y*A=/\׃Woy̟yf;)3tH堿P_`uA=ݝkNt]nb}׬擸*![sAR22* y&ocFrfHJ/r0b0R,2{ĄAF Y,|п<|efIaW*QjUJ扜e2R>?cz4ĒC^ ;ֱSݚN įz*Ή?˻m8ӭIyR2c?7;ORSr2ךd *u?Uiz Mb~駹⤋yTY֒\ݭF&NR;XWn'YiϢTE@B> HAs.ǤJz Arfl! 0]!j108Gfoh*s/2Jz.Y !1՝rWUZRuiӝ`3Y⛻sy-ݶ?b|7ƭWgxSJN#QeVK)C3)ׄτ)]O+'voʰݜW~&p\fw^f6ym3K{\Bhڀj|K#S!c<ˀV s0H2E3tVB/=41P-g3EFYV TݞcvEAޒFHê{7Mr ҽW?~7?оl[a=-sxKͱd ־c ztar+ծw]/9|Xjmje!gBZqcvH` lU}u1:kA\g9kudy^#WV:,Q`:PCW;-`-d%iW蘼3YLe(Ud:gU UKKʜ:zq/f5BIw]Pb4DB \\x7Eh3A=vj`\2iq{?Fܕj"ԁc-eۅRtnGk=H;nwEi\R:X_]>}-Q7?/χ2$ k PJ-&NsmH_!tR3 0`` k"D]xmr$R<Ÿ IJz~q]L"RM\Q!*yj{T&sCmbiⰊM} ip}9f|9[r hc'6\Q9eȦ[eK u8|&9p~8ƒx#<|&(V?|rlbY2|74˘B@Gx[Hi<j#W<3tȹ/|#'ɕK)J{L ̮-W*O6)gBTy3')^\'t,=]' ~y|׃{vwdqu4:[T .א/$E:Gt2n0hWm= !NOOھ{( [ă>#pk*c 1;rDThyy~vwwmAta1ageUϼ\\ztvm*dE\K+fu/;?<]_[l4^Xߘ\^\Sk|o_$ZoH |cDW #,ߘzv~`7?މolx/gųxva^ OJ@OF%N6M3Z즙%YU3-K&q\'lü6)?2V9J)5nqd%8m@}SϰdCy6:'^Φ<3lvtgLldw=C8d6ܴ^\vԋsGgbvU=kɿ`?ɿ~D lKoL?VH9ɿg7ÊgMe%ntFy{- 'o$OcO-|_?b7/֏+D `=Wm/s?q~'պV'^MRrʮ4pmP%r/5f ]3}zm>Wb[Zm{l~"?~uDiyy͏὿hGפrPrҴ(?t|f1bdm6ϰI}ԖIgr[>7II co䝓DϮnzw{/y:^^\읍nx8vߐ"vGK:Kķ-9Xmak?苧67ӝ첓y&jW./nb=E xលs2,!U $[+#eb>ABx\&v?5+Μ-8r7GִACIV]5&k#9r:]wvE'Kx+o7@M&jZ%T^S>xK!j 3*]6<|4o{ OӗMd :Ynm[ q;*p /g,:?jPF5%Xwrwc]߼Uo4<+P!>/w؊bPDtvM蓛kn#(^ +ܪ Jyܡ0fE=Fr#F5X{Uo< gkP\S:ʦTjfМs3;k VnV4HF5kՃ*M3| zğm:3BBvҵuV{ *Mf\_?4 o\ ֫bV95F O (|uժ hVV'4zȕŪ Mčrd.OkC`7k̦rUriD,0r(:ff-` >$#4bٹ&Pk|OkO]GEGFgAA56BOo;GgsI/ -Ӗq)!-gќ;?.IҵJBɺdzfOzU U;$@ g1>{%!Fh\VB $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B @*sH$l@[ ٓ@2X!F(!HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !zI>$hT:(gOY @^G@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B}$7Ha0!̇B k==whX@_# D"$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@_ ~Ȇ7Zdl5^mow7o6]Z}rzur%¯HnˣaU+dqNj>c]g#fNy)W'm[lr'7/ncudGbYS{wzm/OrqY ;|7}/,5~ *bx;Wv+&VU U._5|z-4nlo~^G'יK/l>qQZ9ck-|ܛByzwMGzlo҅wfzLK/J(8(#86ddtN An=EDhct7`B8ϲ1MI?M t 6&Q89$EƑ"TRXSEE1o۷`Գdl%Fe,%+g 웗o`폫C `]¤`!`x sZD)Z0la% ,ӝU*Vz6X#vDtIT\`0`{-MGjm:ȵYQC2lKSBH0:/J<|%իѾw3AG ّ?0d@5 44QWpHR˪H/r4HQi#AY m0bDEم;+ܙ3F;S|j\gB/5 :H:Hf&! %$G䉺a_xJDG}dslӛ4E+fRzQ`cpIAbyQVvh5zv@QAsglrD`p'Q_d2spl8h4-ýW_)qMQYβ|!Rahvrt7,b1_o"aGFC.:M L ?y1+wq9`pGz} ,d)sdԔ3l230f&RBtJ2)3ǰ+ 9 vs,in=WyXMk3k fL$(yRx i=feK!N  V /CN"E"$Z~\ `xFՄwKC Q 9lBhHQ@$V#Dz X/;-IhFһΪeꮥDqbs7uZˈnQ?y2^XFF0>2pS=+"< SG\4JHwdI&+߉VIoM`@:rB l<#ck-%11nj̛Nd< 7HthNN@TI9ӕV,;~1rĺqccf]-<+6@QnˮocR!n(ʑo\;NF( S"*%,I `"Jʃ4QDs; n|[Ͳ< zjsQͷRZRKU^_" qs\U02E]ī|o ګsÆ)Fٽ^3|%m FIU.);C{-s mg{z88EQ=BB&n[Rqdv5.pق(lQ[jѥB7R"lq~ HI˂  =*D3m5P[;:L\v8#:?J3yM`x80F!<_DYvc5!HL)A¤wcֆR]ɯL=Bw9r&C;( %JhU:9VKITXdqtVЈ! u[i:+x!lI[MԬƌL"^ˈiD-hnDdنQdRD\W|"H`-[ K)H< P RS r үvGE3XWaDʩR ˽)Q g;I" 0Iyf:/ 3gGSp(Ӑ۷9hrZekܚ\Te.snC[ZJgD?ZΣ S@lQ#!HT-*h-J@H1\ >jK ^Ȍ ^`A Ɯ&Y2fvɘ˔EBYNɢ6!6wmt֠0>%6rgj{8_(An~j!^܇ds1~ K"JW=|,R4=k $8ꧺJD'e=̏>rTc@)+ )MJIdtd`+ic[.{ peY;HDg2(ZMnxW{ؽsǎI1~$j.*'YP(YW kMV,0 >i8Bfd(M=G dC8Bf8&}W/Ma/-c<؛~W=y9h<" 2 Txcl[RN1N9XaiPB80[Y:U9ZсMGl:4ph__$~ՑlwLkiɡ~Q36:%` uB^Ԅ6/:41PrXn7uю/MܗvU@pa v9?61ۗ@vi]IpCg~tx=0^)DHtcFv-?mc<`mT6!a!9n:{X:*8":*8DH*ԽqhcM0&;V*[*5mA%e=1cO]}zQ!%xJ?}k;?a>0!p\v A۔䣰0F.<&V|uojvc~Dez5GM4gȻ!vLGg\+#\B 1IX $-1d0p!+bX"ey\(^.+J1'8dDrq$D 4$I;SS@I)PɶJtHZ 9UddgM4ee;7q8>OqaɷWA6O[Z*MGͶ颣T޲%CUg@L.|ywCuR\5OXBaheg{bSd%YH1\2yBJ#Jmc`ՋvveQbC%lFIK "͈@CQi4:gs/@i弐$ԃ T1+$/KIF`A"]$-=Yoig] 0eջʗ?-ctJR]yʒè4)[1IwLL/Ez?PxzF6$*J "`cvZAt}>IC2#'%BΆ'GiIWwABX(*`nL0.ytbp8"K'D+|?gdG }PD&bJҬ:t~E_5 <`4|hOON׿r/.CFjC4#ty%[F8D*:,5/Ь1U$7h1k4jfke^s׹<}= DpFͯ+}qXnΛ~uY]}Fw0L\3~va jnW-ƛy#;MfU^3έN!+K1zaSYeGa6xv48bD]\[;> %[s(;7€_ ߲Yw`Jʈ鞯ߧ̈́^we,Ebs˯tע.{d/+Yj=5J A104A_;VH6&,of/o;թ:\ܞfzϽͬ֍#.5 onJeGOZܽ`E]n.kmmmr.pI$\^-om pJ ۔yn[:E)`pEe9 HBҁa HYͿgMBhe]GgI[T@"TR#dbQkK.=2ë}&ΞGgߟЏ$>rK:4Jl<l㌘g?zYȤTcoh(:magY#nhK$Ib IJ`a#*FG"R&J+ʡ,!qIy]u2=*Ƽ6N] c^BpeH w!+0kV^)u.)*7ǰ EF~}=CmC[Jw hiM#kW.{F6j'Bk8Hh"EF5&߃6xt$ˣ=#mYPX;Qß!˶R";oHII*rXoicMxXSfD8߷FzmmkRc8O XgYqF'Bひψkz !df$#Wl_YpnU]H%Z㙜- Fu:r1`HJDgQIqztgs,/E>m J7fiz1 tBm6MKPkp6b1T*ƻ.`[{zw;{w<=fhv -߹7xƝj:%3Io%eeCi8N6 <`][Ocug_޺*ҖBש}OY u9TB,m?Sl^k0I3c&SL $SkERC{+pO*O9jufHZ{Q Lj"9rǨ^{rJExjLF:z/W_ʛ9}q嵼bI 2!EW)LZ)Lp J`RN5P<zmM@bB?zھ}xrI%X Lpl#3$)!ӶZz8 P{g\ TɐupuGfAV9"u>3YŸZt`狪X Y ^vli&gka|)Yh@Yo59g, 1i!kH1!gGh$>lsE(,e?{k\{œW"Y@6ʯV:-(77VW^u3}#\Q+<0_O UiEw͓0[Q԰^7/G1XwxZs&I̢ZR;Yז?4Y%( wAY/.aKi Hmpw.zGhGoxx1T97ye&OznkOŶnn[T3MXgrg>lUݽ.r[۞!fu6R;>^͌*GU] >#L+Xר~GuycrƗ/YOϾ_ξ3>;x9YQGG G^[vѵQ}>b[}|[kzЎjn H͏7o俌g$eVjQ7[ֳ&G &/b_%7[Z4\*=+d qN;cwI$|#@E j,=UQ{KʦcH.$~Xp KwŻz&A J'ͯgkC,f0ya zg}^uRo:T]:dw8`~ͻjH1uLUoI"!VS $άQ(n,x@9b"0IpOL!hX9؍) op/G3AŜ ].e.OWy;!Pp,! "ͧ.@kN!Q.HR3`)kLH%9 DxL%-x"!ule쬯gyO;4OѸwvffu.4kYg'^7`e&X*sŀ '['iB@IF4 ᆔSiHN;kdDDvR9XV[\SI3Ew1%Tɓԋua ^=Ωi]_VW4~EWW3_]}gig|pW?r3Ώ:Hwy3v;_c&eozZՆ귻/ޭɘ?pU?9@IRsPo9PrI"L-+IeMwmN@;1fS@\?=îm0AyMG{G"xgʦCUOxMIqVuJp*EǼe^`N0]O"Exe0t9 7w,4>I#jBT`T% DXeR&('~6j|khT(2g85=,^ZRX85:Ҹ6(f:j;?i^FR~~$K; ˱!BRprNo.zz&g2F?lKQA$.y; CLb,wIETIUPٚU>G\JqUnq叼E=SUk'!砽[N׶c F}N!r6"gUIqmԠ]7|Np@W|ɣwwUcRSd ΢B^gYsJH*y8fj)^s\s̎~^̢B!UB;d}k{B?:!?pjONN?.5|Wq1~8ѩ+;%mh;)s@M"j-u)V3qRe/~-v]ǟw埛<<N~?χޱ D_$N I\>Amj>ȎLx1϶_;qDU!j6Tlw#|V8"״:w<ȶ5sQŶ-`٬ahaqPfU ys7u'wm]j*LF7x )oļOfb֬k^mXqϖGKtDӃtdӑv[iݿlֽΌCu'O MBPKtT}dģCY:`ɞE//*Ы׿óߔ)L୐^zDb ͉ hdу{8H1zY}:llEM{3kwj".( Vs~2??Ӓfvϓ S>13r|ռbJZIY]k'I Ąځr5BGhXf tW |$Y0ʨJk+e(JHjʘ.x( qMf:\GPx㩎F&--Xi#1a&A]~!u|\ϭKC }!0wrry!:RUZo1^}&ʈb8^/JٓNq_W,k+u !APOǣ~*fMPz"P%Z3+rv]f7RR (!j$iy6q<دKЂ!ws |>`ٓF"D8LF&p h&YDt.YI1LѵQ} ;Gh5! M.~6sn |7H´z4fQë7fԶ5ⴀWȌfZ,q$ā鼚WګOXg3նem-NH73qᥱM jb0&F\3 @19aO5Zks c*XwF>!$BZ;{}by4{:tM>}0EC0B:р 8IuJ|:*LՄW%P[ fHG.T/f;^vrVř+ IjൠBҔ0g)E,9ǩ$P`a)s!` [0H'|R(ס7G1 b^bƎ )Ag-QVS3/E_,vɼ6{ 3إSzLdj˱4$PY+ AG8JKѠ74s:OA"AR+@س䤎GvØb p`ua$mހ5 -zKq-9e'/}4v{{h2V?>zsRf\).IC)()NzZ$͕&$/w.:LI[Pv[Ls(j0+1@ V'!rF FS"h@gRZ0(s2M9W1Τ!9R 1KL~ {j"'zV;YW5@## ti':d.6 &Ԏ;I4',)HzR5gaG@aEd,HuƊ= ,0GAk )j!I]J(0VtP~'u~By#0(Fl#fgs 9J1) -Ig8,&[q9_!n@52~oM(Μ$b-Q8uAד4YQGuVDv>uNfmV[MP,zf6DgDį9&kQK< &zFmDGԣz,X퇣}юbvRV;A'yc lR7Me>vtmI$r< 6y` kiCL8yWV;nb d^Tor@[ e5ĵ]h&~SVZ/գIoPkR6UIx/O>c`Pq<㣜PaS\ |+lsXMF#W2l1K=cTfvK.5 /_ q{fs߭fS ꝏRT޼fYrbc*l{T-j2xxx2Zo_A KH!pNL09N >u :x6S89IN㝜",B 5c)RjљO5W`9~7L>=]ϭ8mj-n1(rSKVӹ ep~It9kdmlHu&)xvZYM[]ٵ̭EބuD3J3Oqu[Bk&9a,|WbV%Ugj{㸑_iP^E!^|dCKQXQ4{^$Yi<=l9d5UX-ic] }Tyvw1{獡ސ"Z1à@UDz j db$ljXU tvhd$CEKBѹڶ+RyQP,\Ba9UXnQ-+i6 d/ʪhnD.12Rʪ@N(2QX(@1xl;l€fo&^{Vh܃c`ֽs<@ ۟Jfg& gXxڟ_6[ށZ`o} ), vi߾\(d[(&~.Z wѫd\(*DYRDlI6,E1(=QFSKcH"rP.YzFWO4V&z{PdYg59~g(. Qd&X-0*ϲ*C+90sqdx%1VOTDfPjD#]9L9kį Y-?ί*:ٴ:4.8 2&c{8'ɋ2o.i( ?],)=J#{ON+d 1IHƓG*2)-)J"Wm:7Pr.1㘪bCʅ\|T9PRLN:& %c3rtfTbcY;YpIQ}˪879I^Mtddk6f-RR(΢DU,rBJ`RR{&Ya+iUҨi` bS;HDg2(0ba+Hp"q'|bnjul*}cwR{>*Y%Y9sb(+KT|5&+U0bj2k0NeM&&NJ#(b.6Ġ >֌ŧ`l}l$+dڲDIU+lLVSYzcl-BN)e@pvZ4 ccFK_HԩRL1!GN&L4 M'rKfQUG~\dlJ6\;b#S^8U5 u(04֠[m^KJSŭAέVǦP5{<"IN=r4_[8faCApg~d#yqpoQJ 01?WI@HtcG=?c<`mT6] BE6i"K+ |FbJkshQl"ɺ->clOrmJ'ލx>XѨ[ (xbAX`>AmdrQX QZ .'T.2*l2ׄ dP38H2V%RʎyZ崆Iε/+w)hU|Fd_qu_/uꊑso%`Ynag®u[i%eg(*Vie%SoEr_@q%GMl3(+tR9 JUɑ+1^{rJL#FW|MNu?wrD de""Hk ҁ/21@_gϵ5H&gka|-YX :frjV!$c̤-#m +{'<A! #O1\Iت,e?{kpZbūZea1H2ON{]剺I\*?YZ{jbH?<31(Poq5UK]tC&]*tN,IoCzWUfD߾i?wːz w3Gm7mK<&NgZ} QF()jLZS:'W tot6 |^@5$~rA)$7_`UD!toIȔ 9xGWY|xf_/7HW.Nn^NV+rGhV]vN]^q\OD^R1Lݬy5񟃙;? 78_xsS9PpjOq9V빽a 6 ћ0^N _!VʑUèh%WssČi<d`źO'W^9:]|ɱS+:^u6nwitvqt?}?~ORys3pA7U$(I"`mC o~CL 3;NVn&֛xw^|zzZ yP^ӄ\Y+7 ʁ"*6`bɈ57y$*}I" `*Cvk},EKXx:@F`* l5#'} ij4+c&LDXTLEIZ&,$KNV;}7NvA 㮧/yQ>u{kEvxioӮ"'g~M&N o]ow'謓SΟqvޮ^do{FSplr2JW76y~zi8㏫:^D~_Satuoy @׿ Iw9,dy_~>nq;igj=6!qL;bZMduwaoìń [ ter*x2I7~oթ7rۿ͵a =}2JvBPuB-^ǷrpNZ֩:' QVy’eqan8>sr.Y)?CnM#9}c[vx> .$7F_bX:uU96摁KVaHzۚn~N 5x.[UޮYn1{k|L64KҠrD=jK׬pd@dNF'jҁ  ?gMBhen_9X| E Fl,㌘'թ(IEh;{x H RuDQA=~/HM"WCYB`">'.v;ykD5yKH2 d."d%f͊+e%EBv32R$3m3m,N A@ 9kl@ sAW!( sйmL2y\)!) (cW$e1#& 3M17#SdgNfʲL!URM"hrIO0$MBYk1f^D#w̲6m $bdGE+Ȏht3+vZzQHVʘX29Nk}!x_W@_C}%|ٶ/@wgv1YavanVG`^N#ZLH52=|-Cﵵ J9E,(tW3}febㆴ i)Jkz !dvWl_YpD0/.Mb69[ (:2@1U*P %2 5 5kFΚl^h[Ż3 gM>fz uJhD>HtрKP>R:޽[zxں~{NmZ:!j>4wznxUJ0Y-^->l~ud\ֳ,yh#@j OXe壧~\_4]׊Eou%vu%vNYE4%cJbo}$ }tƚE}}]+ol׆"g6)]Yo#v+@6j9M`$\q00U"j\F9U\DIMhvt,} -P‚v9!s *!Ȥ8(8V\r)\ІH'  h R"NĹ>!J{D?6-17,&U}`MM~"u=<Vgk8xmV(Wa7~CǻX e1' !֩ Ekzu|l{ͳX n-Yvj5.O`#?K%ìm/Z>$b0u-YsZr^5tЮöCZ/@(p8CY|U^'$”᣼huA KH!)pLL09u(ԵXpsKkوV09S`R b; D\)Q P)1Z:Z׸/fmwzm?V~soc6`1r-C=N>T:\$1\\~j#=T^TrN[(=ġ|=㶺Cg}y=~>,Xlȭ#<նaa8[YVwC3g1䃪y6ѣŘxqQg7mmP//Gٱgk$,P}>d3gSZ }TPS:[Yj.w}xۿSݏ~_O>P?#:'sm"0d$8kC/Z547kch`mtWy[]_>B.46h{~~ U[oԔ&9ز85*eЕ*a1ik젍T5vPq@5jPRfMLVίhmrb h u>j.J\gǓc)ޚN[G=6e%_.pVu;/0;|g;x"[HL*H(b%UAҥ5D.Đ&q<{ׯc7ٵ) (9!?e BMN8eb1ޜֽ"P|7Uh!.G!BjPC4/ck6ܘ]7uyEa ϻ?Ra΋d,zw#k=|c)8I':8)S &6'fѠ%GVht)3eFR왂M"ObWX(f/El> c;[%ɲ16_.'Zuegj(CfrzpɆ QzARˆbkt:3ݲ:Xr!?`8-:)luoUOja]t*"ZYY8QFBYJ]E:E^ë c=R"zON ;hneF>Zh13Fr(L+iLsmژ<kzf-QAQ.Z|ZE^̓h%4 !X)NkꍲVGBcHAic ׄH4u+ :hWej5q@:q.4)rĄ'3<^9xMf7%!tXX^)dL QG< ʣQy (GYok2%eMS}%}59/A_n{s̽cm޶=Dqe6L|S4\RXƤv&)+~1}Ύňr1k;d|^ar`kG)I9rk6p̯Fpʤ4PE$7'v`tUȪ(Ebhi;4 @ܘ&)jCKRʱmƜ>Tuٖj} V X$9w)z\>/M w<9zTl2~vφo TY\_ my1P44Kg)EAV|clSGx^LeP{&ԩrȀ2iuak}S,>&M .81,BL* :z8a%ͤtSؗg{< gOUT?qZWW12|S_Ǔʰa6mf>4G:ws]Tdꕨvj]Ʃ'/r.{t8 }U^>-$`Ϯ)^[ȝjOXym鵻G[?UEKl]n]6wwYt:={6e-˩]y6я=z^j[fWw;YvK6?)GR3<.\5g?kzsئ\m D]έ}xk}E7Hƪ>p>E"VƗB\S B4VEmK'A@u'ӥGy%r3E ~‚v9!s *!Ȥ8(8VEZS 1PKwQO$ZʋPHR gsSQwB]W;h% ȅV()5-&3B5x`OZ`u`v3i7V8moE|\~,n:mnz/4ٳIc HM(t4+p2u@/$8;/D,ѣ=nK=t֗W0//?{F YހC:9[9[fT7.b䒔/2e̅$JYptt2l<X `MWhsPmoؼ$%7; {wg*]-SiOR{4Ju로{~Zd.l'2fw^8ߙ#sR(W9Sb:"Lh.rR^ÔM&2*'u,kIA+͊):N6Q{{ ;'YV޼_7q,bn[=oɾހ~@ɼJϬ!Kvf8 M SΠT' $[5/ªOm'j fDF3o3Yl{PH\$2Tl)1JŨ$TzC11-spE$>E_KkbԬ;(tL [oiw3Vןx8iX{LG+hvx Ɛ+叏 :!uF ӡ3SLz'>p, K{* *(gKc RR(bB2zT@FB ԃ*+ۑ_i4Oͷ̨ٗ$SXX/Wb>48> 9O1/l__N- ~::\$zui=<7hjؗ\4ΩIZ0Y2}TR":&Dӥ)|xgN(gzՉ)4Bg1ux5?_{}Jv;+VؔAi%խĪ^#9  y+g)G$N o0&DS AB0d\AY9؈C,|(id*Q`c n֝|/zz\PmoνX]}ˊ޲|&i߮uxFˢZؑU 2Y'v: sl`2*{&^}Ԏ7>%F`JǛWN|^$dn9%_uعX>W yլ:S˯=ĉj|+ժz[U7W}*~UF.SJP,T)`My/Mm~<7tl0b's߸ӺCft|9^+5X=$t1WohA?O?ݹ |.~ !x@`U""9N遰< "k_]Gp:o9z^fi1! PvI?s3?8??GzFK>-<.O%'(oq<%H;3JY S":y ŗ K|π0m?o~K}t>;ձl:o~87ܩ7.j.|ű7gmZou2y/"iټ_h83o6`͏lP^JL--iyu}0ʬuǫ8PuVҪMwc93+n vGVFWfٞknxw;C%p:ubcV77Cs1n+I}?[\CfiaLNX=WG]L)8=<$ߢƌE&"*X 29.8s &"AUBd3*Gّyhkw<OXr4X*1Ѻ[Ep QVh nCƹs&W%0nW?~[ 4}rr3ٸ,|;N dQ\.kVC^xXujw|WR KG#?S42g]=l>̪#hʁAByr3" ,5RÑ'GOQ>Us J) hJBѹJd6;)R& B:5 IzŚOR6(}HR& ( Р VdYwcʢjɚmEӻmT=x^)|k }R& ̗kgAY#UVT5?uܑU1z˜|age-0z֬;Գ^}|j9!5$ tAI~.dNdعW3nV &P0(('G9/߅?ނ^$ֿb''}RJX>6SS2)FH#Zw5W4mΈ,CK.clN-Z'嚃UN0aQbmM,[_vEծXwot=%AiҠ"dxuv1uݬD"M  >C= BA=:>>5*=$ ʡek%gZ 9;[ےe9j Ҙ,U)(R$0o,՘<6}LYmD)|Ld@&%I ^Z:T .%d lȬm51k@ 10~me}֕90˱{#DƊڬ; PEF)nQіTSdtJJH0aIbɱT9R BNN:RZHcdl֝؎!J2B }&+2-N˓u8ζA}-[DR(ANʚbJSSVJI$51am>0Tcm[uSҵ!gm'|(aWb."oͺiQ#6|=sun,7AIɹ/T\wib*0(\` +"} Brۅ:P;TCD dL$e8UbpVl`AJ-4&A*!R `M a.P۟`>}4:[e XYsNX-עR.Sts&!â ^Ӡ9qa_P9Lr꾁nt~%x8ߺnze6 *p xbvzYa:6 n?J4,Sy %K!>*[p!Uaa4`22v@d->ՆeL=3>O13yt=__H#V_^ek38@3h@+4a^n 5njR>:l"n13y|j (1]@5x&?]xG]8/7[ށ:OOWmX^ف^ɿ~aR#D/AOF=G^dpZ N0rQ<hCU4XbA1"J!|{Q3k wZ&0ו`,j3g~D|[0}8i Kܩn]eߖy?1< ǢDch;erbYWmgٗ?6fe =1iV gcsKis_PiLL0! x.1MA#|_Y@j<_iun2%H_rO G$s*D>pM"a @yf>+Mtirr8N@7P9`J /Ҹ^FJTa)US} J\9oTpXq(sQσa&JN6{,RG3A7TFŁsgՎ1eT`-il8'K:UHhAܭ[s+JN(Rb@ )ǒ`m C@C3&v9]g֓g2x‘*%X ) H>?ζ}whjT/Teax\ec"Т"vCu.D;-A.L & n+ae2^ΗV9&xيSɁ7{٘+ qp|2*s,SQFey:Y_{6Ā,w1~AWb$U=۫8hR# o=]/N0g]'Sǿ +.aJ {ݘi4z<{.ke}tMupp1r-jJs:23] Zaf;YNpWdʞXnn%+%C*+W1K^^rg=X39F^)^k=U+`C'9z"{$w,|~+GŰPvAQxĬiTM1jF}ux׷g_z:틳?yg곗xy :A0V {{ `Jw,tеjkiAל69oүlr+]>BN)}윟j1v~~~ ڥS_GU@IcJMЊtev (?a6DMTiaTL  р!PU~͏ORkuhNz@=&s=I1I&j-LeՁh1.EkH9/6{ãjWT_hb\y\ x&}6gWi5Aj'3"i|^~F:;UѤ1F0o˛MzR&GEPQьhÄL,a T*p(J`.KA\Dy y`,v'g|<414t6a %K1T;°!oyL`U0/AZe2. A‚xXN5h`4hacappĘq6'!K`{T8axU4)"A & $Etrwp Y.Qrlyyb}P@G {Tt'&xVx3U}3T^tP\|b9OzxH܍)L:|>~?#99E.X"5|6*'0\fGT~׭u^fo!.u>1 ;K&T/|l]g;c;-Nh{LZ4SOP2$10bsr| -6>hpo:$ҁsgfdug:㹼Wܚn&ٌ6z$G}obAE:2CV7u%+VʇӚE+'VtPk-۪sk-Ww ԵcGL6hP'|A}%~BjKgN.%s"1ZPf_8 3 H+6/(w):O;E0k/4ҤsHD.€)sL*'9C$`3ݱ3{xGlϮG*g6 .a6U6M"ea s[}n1øĉJ/Cw(gI F:V`4jC,$rvh(rd7V|{;.l ښ8ϖP^'1F{ʱFC޳(|p*8( Q4 {5Wl˱|޳Ȝ&Po}rJQF3+32DRaM5 ȴDLnGmYO*(RpY /\HUҲA;z䒦!B1iMQ91o1&1i%«c oƆs 7Mf TLMk|EFpʞ[d>g()'(,a9AkVj&@v3s hRR 69,mZll8㒙Xu?gSq΂-X?`~TP6W>EoU0O <XEGòT}-sqEB %lW}L%cA`{ጵ"Eo3Мvy]h,^MIf4UC9_Ok/p)O} GguaB, AtF6/_;dP(qRjz {NM <EPyC;j|<vp4}%+߽Vw^7=6\Y N,aŕͪs5Po |Z+ <]rv[5:SH>LKLzzPCR3gbAɜ,SR QF, јr .kDWu 71IE 3p#2eeE+8wOO N?MpfoyS8"W{ύd,]Ыyn>yIαY՝d1AF 4[ȒYH$bV(*!sϨyfNhMgUw^N5ӓ7oW"wK:gT{VPJs go40nx8)3nOF OQ&ndE2:0j^DTc:OՠGx0*$*L`J#e/U)G"wY9#R* a,+ˮ^ke$d{! F&sm`TMu/K鑧GmƆ7c=Vmخ?>zqQ3m}3A@'RS'59yRc<4EgРMQu"c:ou|W8> A3Ha)k*;M̖31y9O{ɃE@͊GMT!Ҹ)ioi>T/$SXZOgnb>TY{03odv}FP5i2?7)7H~vs7a;s#e?!F9fcZl"W/#cA`%xiϭzfqj{>igBlvػShγCBllr4-v+.3 WܗHdVށ1*&,Ř9 Q:VՂ*{8'ɩ]l9!7FD2VjmW`8Žk_]ik1DqEY>Sd),Y S'YԈ#3x Nl@i8d)$l xu~q&귺c8Rp4xK'/$^;Qa`)P LNh/ul XHx7o#nnW?SμzḍȗS}:Zz>ÍH*p>+sm|$ޔE@,9 !I䍎Kva+Ec:D< @o YXh3Q$SKFx[R0mGy=}fY呞 ɇ{;8,խI[|\g_4[^Qd_G !=Lߡ4&kr2HKjv^Vr<(vgxJ?=5ӺfR͊!TWUZWEڸWUwLMa)S(S,7߯XMw{ez܈FaC'pȍ{BΩ1y48Ǔ̞43F }OS3MFŷ ٔ4!e݀Gr^4(2֦f#!{LQ?OFzv. OLYfz¬+0)M!#Nohן;v*~X4MҠ:?! ҐMۻKS}5-fh3ulo%}w?=h2>Zy{GGðh0A?{1~im/\^{w#YS{}5.>KwIewƓ+oHlh6gTک:Wlsظ4+e(jk)aц[`7ʨLj,#x&vVQig_V`ݩ4#]1-1;GThĒKVQ&s4"i6c}z [*;1jx}J /lS~ >6oS89DJQYd:Z2)ь%h]vN+a3N6+ޯr5X[f[^+}c#ɰ=1_K=Y$bskRq3CRzh=9di<`3NMz5\l<萸%?UBn+v璒%% , d,juv.x^q4ge>e q]F"6X%,̽*TY)R/CjgXw)9<NNsԞIۛcXh4.>4}ks4Z #{44A2sL.ْ0gK$v[rw.YbGshR%s`\12#G$4'5IQAZNu\!v"'nx +GǤAdelSZ$:YΪsO;k+09`}O`Kʔd ݛ$\FUI kf%MB L UN śk~fW g8i~Ή(" ρ@o5. 8TԪ¤I]0ݦE#482YlHdq1\\H(*-tpVqq;EEJT4fn|aZ&e4ҡH EVHR[r7CiQԭ})Z凫$Hd̐?-*"bQH\31e4rJ9F6LRvA$<`2HϹV> 2&Öq82\VƮl mO}ӕ{(xKQ`g{|mZ?O~bgyisl] Jg{ H:@\[)̊nCqȷeB,F*@|;f]d639bf{0qa:I\̽AC/gUȓ283Igd \IGEFAj6K+ rU:J5 (l4F>T#|a{;0El}d7tY˫E|6g TRxtM[FȅK#ev3f@oZI.I: )DР H4U!f%W.vkF}uJv|`ȫ]v^H#s9X %;>iKѭJI'4!!]| 7PcW{(;*yMߒz#2aј?PٗQ%(Տ0i~AzWW'؂r Q*3ߍ(&0lKFU!I^iu:Z4J7QN=-QQE!$fKn\xL*cek v捼og]-gUnSyj/SyjS5rT=UPS !+7O<)Sy*7OT+m4Nol&KINX 2Gڃ NKe딎@E@½C_PpOqR$"þRߟ1˟C- ":&::9=)'">xsw6 S #TGo!Y%/-ыGor)Ih/M2% n.m/h8+ODNTr7]fyz џ}#?_Vo\shǣӳ录^ R(|rO67wm39u&|4b4R\_,IN;XbɧFwWsG{.N :`ϺmnV†nfJ:R&ִ|<2# >⇼.s7X?A ,M88x^O?P/Ǔx}…=y?N^=:tic8&0-?J)?B[SO05 55ͷZZW71!/y˼G;[{iQKr]g=Jj$xT@W1דԤGHlR[.ϷweyM% VCB=n#&>OROMI A, k5zfM೰;E۹2dлh3tc>yamwbl1D7jTF^2_#W5X-1mΨ$Wm[_ S.ٝ/޴~VҝScp4;Wauі*€K,· lkBN֖Py+ ybg'YE1@?n~>աf}9ŮQͬ<DJLoD'u%~(}&=OIZ_v/AtҞbb4dqO0ݿ `Ua ={r펛,G73Ft;3߳RJN>9 SiLNC\D>hz5;)bZ%'%gJZ򣆸/8fF^̦ؖsGJrX~5Ot7'K} ^ZEWN#9.ű}J nEӁu+2ڕʭ|KڋMg+09`@' p]?X$prST2*Xj-c7@4PKz$2&!03Su}暢wAH0mm@NLr Iu,m4({fjf"2mrV@yfz,ޤ}v:q$O)zU+`SǜTeYڑ $3BDYuB$S0UHfg+r]{Ə?2gv$Ρ,gʢR&֭mʞٻ(Z‰*rZZ"ٳ/QxxؾdvYdP]@g7Xt̘;#Y2N2y m XӊQx \ndc,:"2c9d+RJ^ak uG[TEdqL9uI˘!/s[M`ڨ#$ih6PT$442Jr:2/IEh?8J;n%mrP lUP1| _xc*U2dhce"`iqZ },s EJ'+yK.x[V@[J}|~o_Ig{TsFY<G+:w3gBL VI['nؐ]V΄9o+qB;2b0"mݑ1ryM1 h1(`9Ƀ:(ht`yISrWPUf)CN@~w4 =%Jr?;O2fiÊN0\N'R# n\ vWݬM%+/X뒠07}WLBbDNѤ9ZtnT!뚝esE v8't2V0Zek!68uN!^p"9()mo+0Yݖ|:i,TK2o{_{wL@-61kfOb5+WQ}7*{?cW_! Mg7æ,??4*E-M;?fY{v{29@H$5`cʊyBj8-r }i3mk" GYԶ5 ᦌTM^S1 1*&\Ey2ĝSRwFM9ʷeZ1wAl"&$3&HPY'-u5lmj9oeH?'/߀fXp.UI(\_`-Y! -::b 3tӇ!ixX`8AM`I1HH!;զ)EF_pQgæ sh1._ N?w!zLc$Z8xs{fL5oߝ7-He/<Ţ72 g_nUdy !Q(߆A. {Q7ӹy?57d5޺FY$kJ>*i0덚5ݚ`=hkA.Ua"=!-PyP`ǃ^R BeGZ8Γ N8PHka '=6 gP:ϩSQB]W3]zXFВ+KU4&3k:%"xbAsX -jkRW^2i[?Kfhjz{zi]8ffJ(&svՄ?)#νFĄN[-ۑ?{o/4_a%#< d^bbTq+ IjBSg ɹc %mQ`#(84Z(CYFe+QxSG<'gzje[-lb-!&9Ƙ?Sf|kdsr .9()i@NńL׎dF ?[y e%PFi꽒8?&\JT BHr WU#V]vhFh^NΦe^Nί/:O_5I%U`){ðfIgRA L59rB(Ifp+" !W,qjk%N(^2 *yl^5g}8gHuvϞjMGӌzEYId6z=9^$ڄ, uZąTÐ&&ꑏ2tcq|p~rѠu@m<1E"HN6drTk%FCc2rwmv%z.R(}{>z]sx3+YwK|1DKmSs9NNW4q \>1ܩlOzKFA0 Aq#-G Gȴ|7bQZsf(I#jU,֙K2EΡLh@5rk9z:Sz5$϶H;jf%O8\ߢϵzG\>Aʂ%T9BڨUyRxړ 1hrZSzZqH:cBjIl[JC#å`6Z6zrF8O2wĝjY 83x^aD\tT{iV6GXaax=u]ͭ[zk2Y&CC5Aiͮo F6`yWVhPr|q'{zRI1Iĝ Y唠9b!TƇ" ,`N*ß훗0fױ9]OB)|J4F`nπhzg;ό#O?ﻲwş7.(?'Wg/ĂH~]-VF) ƀ"J#tՔux賲DٰK ?Vo~&?U7׿$?팆`2雗gcAޔ9|7(~TGwa>A8yi$3wnh8])*\m^efSnz~-$ߨ4r1|H7nƇf$C2LG|{il۟5gۧ>L%O!Ys=>8:'~;}wݛm{>7? a1/p揘[.PqE)Ocv2yẘA^ |o{}\fhBk򭛻Wk|~;W|* >D if1ͫ`7B[,j5g>|[n*`8Joڡ)-.R{?Å[&We8}"ƹ9f`)K1Ό`\kf}٪yʃFwp=buw\# P[F4ĕPVs?t\Rw[9G\ JKK/n Ov?\MʠK];mt,ɆU6Ӆ{\.`ܠZr>`цJOP0q2([UYdeܳ7|hfs>XV`iv?`"ф bxb\qœ͂gs㭟G <9H8ZZaIH#IQJ `ς4Xaռ*hdw|`smmȿ|ln#,&&8VH/=S"QfXI !H\I ŽIKL)ЧI~ k̶m,ٖ+i/eu/ = . wc5CcSHyٻKf OFɽEkWMD[g5Nbı qŐ0S6*䮢y .(9? B&Ny)D%嵕2dw$ޤ% q)&B$KM Z{-l)K\ Ot42Gָ-yVAq +!*6֜mN|BplX|Y &'ӻmd&@ Zmi~*U9ǫA'TZbJ}< *"gxw2I~U9yMUv %6ċUłŗ:=կVZm`l;_:}g pXՒ]bȎ~ ?X[JP%TpgK&?>[[ToE } |׋DŽB7 _@zRlvv<ن,L.;>q<u¹45R O{l V}}-v'K*c]^ƒY+ IO,k}և)R-=ofaIzA =Ga[LoUT KbH.{x~>bЯ>ᝳ04Ebu#νFĄN[7[L.vv=m,R*$}%,wmIi zMULZW=$%J)YjJ,ְ_F_tA9.br4rwa7 ;it>l@$af%&5Aͧj[by5=:Iׯ7}rr_}gpqѷǧ]vYhRnlǷG?O!?q Z}$% Gq?lSJJIa)N 8zr4>]M5>|<ڍ[/a ZBKc=ZMy)O>ԟ'\&F 8:j^^ҹee(M>p-E˔ _Xg}ȇi*Wח2WywΔK1:Qt^X@ Ihh<ꭺ&/UUV#v9_h)`ɺ%7[ͩ_M[׺-qVݵț c3V]wiك>KCz=xˌuw-:fwWIRuBˁ;R<9脉t Z]y`ۯ-'1nʻ-|pW.}坚k=H6h|gq^]v߬oq[<,S;LoD(9CD.r!TLw;BG2N{C_x ¬uFͧ 1eJuy2YmBg7c`|}.}k~<AP"8g,("WZd$#0)$O&6sv(2䜈KՂ)HcTLRE$B&\cl:wJ_lO~{kH>f"e"PAdRl.%$Cf-nP,Q5;5mͯRF%sb%J[ `#{6튚kInשeNpZ}Im&'MG}]Oy9w3,Ԭ?텏X&}͍M>u F?IMQvv-iL&Hل( "HR gUT =9ָIJ 0'(ŠIQJk iͦs;2*[I}c! Xwgop{.ݎrhO;;x/t8Oޏ?~[DR(ANZ U)ZSVJIO6jbb-(i)bkm۶B\TklC$Bm;( \1b7Mfb m%Ͻɀڟ Rn,:19s$)}$(C*+UΰI7b-df2C:{ƚېYAtLR*6Futnި_mDV`Dl&Bċ<~Dl[ɛ DWzcl7RNB^DŽVy6 vmR yyvTD-Nz!FY5zij2:t[GMUY}u6lrFZYpSD1c%l"F[phRz]Xέx=k@nUЯz6>)0>ApCE?z/Tka:G'&$1*~ Q.1v Vh-)@HcӽO~*?E7GXBH0ҚX2Y,cT5m 6@Sf"i4D؞ ۣ^VE_W2[/?\w I0D{9;4YkIaGF$(FS s -r8SPddFۚ$eXF jgyr9;Y-W ĬFǟCvxB\.ErTp e<*_EҠ RruwEa"kF>/ (*ʢ;o<"T p".ArB(WYcOţ;XS:$V:$*+η$&eݤlϓ)etkRJ $r*J'=i&Q( ˱C (I256PJ ^-0ZH,8?=uv춤0Bۺ2D\+lTbr,%^:5I\-"r 4E; '){DeO|.˻ lQmAܵ uF]1C:7\-pۃ-!9XiG{K[Z d 2Hz{K /z{k!rzF\ă\a2Q8CFiNQHg+4LjMeD.ry =Z;\ը<.%LM:KkO];sDGM}%餔RIO9,dX<@v΀$(σ#BLQ,7(M(ҙM4&Ƞ#uSsWpY}^XIX"ӘHFȖ8T$pMs ~DY*yA1e !Ԟ/ Ui'ę y+#;"$9x΢ D .*ÿn7hyB_g o/;wfpsOKȗf{TE!CXoK+-1e|Ѿ(x;ݎ?)crS=;1Flx4KH|3GRU%O|‱:zb :=$(P{Yr JA M"V|?X%*>]r7sypkԣҹcz`9#R)ETؔN^jbȁ,1r)riA%D*P*x?XqEmgWq&b@% LkS@̐ ˀ'c hJ5J8P\f B~^wWylp Lp*<`%G0h ;Q8ѽ u+H:S"Z[#@6(cFmy#R(851&r9P&!P}$9c?Xz{l`Z<6.q1eZ¬(` ߥǻ.7PM\*l Fjy&W(A*}*TP@vF &dp甇* 7r(|9+)W ] fr5cbIn 2 'stzFL>x:p>w𩘸z#YENttt8Trȍ&*|8;!O]S_sgB`zR5}\\Ӣu45vS;[|bm F3 q8Jӳfm]횃Ymi˵Pq$E@Hp9=1hhhǠ3r:ĹJ#GP$RߔHRUbkTj{bd ~Q'Ί>Q 'BQY>(GKͿ}WUEjFIU.P71֣oZs|n NR !NJTɍų|) y,n3)\̀ D>Spr)iOu-r.n k kxZ)($Gl[Qg5d?v樑*CfWYhdʝW׼B(aDktښnW{[\;{4׎Gn>;O`C:[[N?U_:Gp@ xBε`! H?Gp퉊*);|"Fg,fZ03|6t84 [o=3^Ir 'Bx CPٚ4:9,蒌T;$rgY.Is5:hFGF5301yLZY'7|}ǘ mQ.hl{z $Ih2#%;'VwSBc҄w B ry3+z\TnRPAj|gv?}򴷷}^]ŝZ2\ތRI8J@TIYB/Ng,i"\nǷdJ.cO%h4ZBm]Mm taP klM:o쇥4QC5$rtj! YOzMzLh3-䗌vG\-]j.pb)U ~FSG&z亦ܕ)aL",=~\-3q5sƖjh4U4jYЂ-̔ܿ븽ܲn mS16` e,l2 ١L6(-!cQlnc0j!SgB`Ka # ! 0[遠}O'oYsզ:0:dyJP[)Rtwܜ A:Gqk#8sԬz[ F%k !Úhݜ|OZ#wR=c_Ahkc0mF\ ҙܒ.`1vѺXm6j!G]цdP둇 -U0rs@Jϔm6XXea\7 ٞuTV@N@NڨP/  lIv.'R.,^@ XT`t[B \ 44#:CIٰhD0QTzs`yUoܔX4yql5M`1WuQ4`ƅ%4<2=K+5Rl{ X$U^Ø VlzTF)WU0XȀkt T{z]ٌ ( v[5:\vAjb4)L@6$ ͢WNKGf: EEt0MUl8n2lFhe&ڕ;F͠j`o]P?;1d a.! "d!v%H  OQU0xձ~fQ D5tSS5qBb9<,:SQ r 6PlڃWb7M-VT4gqJQ.0QQfA*ZeIwB<@/:j<Au+u Jq*BY^uT"z_|} YcI5Nh_*}AY8P,l-@'@Iĺ=(alDj>VqC i<Ό>tK\:&m+j*٠dX1f@TԃL0&$2ȝN6t%ed6Q t56bXU&9KbOJ惥>ZxC $(!4_15lmTn:^)X(`FuM7a9ҶCMZ WB(NCkm8mŢ:"J[ e"~ Eԡ9;sT#Y |If⭇Rk4㥷:p$`lP)4bnG÷h5W jLk9.me_ gqӫc]ڑ6iP\$;v, D4usYpls0{Ϊ2^]fZkIQ3y(Y#p4vf c-냌? zQF6s8n𠗨=a]˩4Ts ^!7Fhoǭ'bE]1١uN M݀u?+;:E'yclZG*ahH 1,*4FBJ'b$Uc`<Tu`h\SFSh pN\7ƖS1ry+kѪVVm0f҃d@&SP VTX~\ےoYo5sPqab2<j7ל@PK}4E:$Ai*`)o"0`lpyP/FVQժip PH՞f((qKކG Гdd~Bi4lpCZtZR- ZM!b!c-ձ SjEP$bJ+eo`j?&lzF^r||s>F򶌒DPk ӧqTg NaU*Mڇ4}JnFƂ"ͳ6f$R?0=pUxv t8''ĠJ p~6N DKO N@_(pz/N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'#E y6N sqqA=u'G8D't0'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q}N BR d38Ȫ-8D'؍'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q r}k{`#jw^w7o6;@KRQ<|je;GʮcΙ攋r}g}{y/հL:B׳Wշ<~o rG7C;8AظZUzg8{wT9qW{g?6Gɪͮ?}ſk6YD9 xL1 4ZAiv>;>藷ĸ|iZ]!`zԍA;8plnZG9G :NtK dv@71QUiJ0L@7CӤghWP(NjNxL&dc.Mh:)Ԧ&tac VN#`(vrjZ jS KNwĢ"I?,}4wad$GbJ<Z~ec}.J)OOp+7C,}z,sYO[vH)w5fh |_͜>f3i8觯ޘ֧~B ({բ~ۮHtWGNLQ0I[W]1Wg|x"oNKߞYptqpp3 kJ?ιqz1wq[e? rP}p#|r># I= >,WIbrͨk;BlwDz?~rj:Jp͞FjUT< 'F떲xm) Wgesx-ct8u>GBl^:[c u?\޿;Wzqt}V\:_.w&!&~·u#ٯ Z~)϶w^K\ V3;H&gZetwBF};w{5hZk~%66/8@ dR,Oi4@ %; H5U-tWꉷ}()S ^vL)Oh84kz7W)&R|᥾Pv{߹< ,?jkBg~Ppګ=W0yGtVsZ2׼>AOHGlu#qϫ<9;%:nf3xNy_wvu}*;K-٩ӃwlK?M|د_sL=L~ XYht~<%1>aw[nDY>c|mSbD|iM1Ku<C**JwcIAᅱK>%Ԑᐲ9% %#%N*PƐPD:/2b9iDuC@tn*@bS1Hq J=K$ZNyxVlUxO\p>AFٻ-xN򳕕mVNʉkh\>,a]x >/'Ƈ*!'H r 3'/C&z <5hdQγԞl b@Se `'nV{ 9ܜu̴{ L+[qVR$t ɤƠE5odV|tSK49iO꼾~p܍p@tmUXۦT5fjش˹^h86hPL[  26hyzN2*{tْ\ێJۻ^=^a-?{;Vu++S>6Iuv+Ic}e~9֟Zuv.ڴwuc&gÁݽU2Zvnl|wx{r7=LF5Ow0kއA g]wt<rք 9 z9 >HVV fM}mۻ6zԓ nJ ঔms͡r: 7ɕDlSx[-ps,ovL6EUNCneS&q")6Irec$rdd}K|Sk叞*|oNKo T"'jK D.XQ,I5рHҴyݺ嵬hfp3<fJ=E#frɜw(D4Ԕ(k; N<ݗ8I EoR1Dۙu|'"u řcQ*Fd%@H461$+=? afʼ,L;8BbEcIGV,8Ѐ`ATDZ`&dH2C딣Q+qa'4cqRæ|k^@ mwtW@!}aiSEUL 61wmHvcu20 IocUY ƅ^%CQfˋDEVR8_οF?$dه9Jw vuۼJzu6~S HSu(+3MyHΖY b +0m`:ܼjV?mJs׾! 0À@G0X 2u4q/=POR($I%K_ EOZEJut>Ԧh'#!]*ItopxrC#{:Fۑhs)6W`L4)9%* W^4QrJ u:a:%h.`{w˸.\t Н BhyНRgm% 1DHϥ$&gR`ʅ^)}0Ow! Axy4"kP*QTCDRϓYLJ!RkuQ#Ct:pHƸT `9냐(vw* P=I)1`}ӇvuV䲦i`ҡ.;x.;bao2NM17mVd }ߜ{GY9 asN*Sg+=|{r'i&''yY\֨pVӿxދ<^4\֕1LۧRrpoROpmД*6'WV:ohlD{==K,مi9Pt ͝X%B'cF@1gz4‚Sѷ#REpx~y5h6!)೏WѥC# >TSoR71 >}>,>P8ߋߗS2 rH# bIRHMAGL ` Xy$RDdYs[}?t ]sxNK-y\然nY= >@Nԩk[gt쵲2aDD\X&HSE57V"FMsFV;|t]pr>r :a춌v5g2 y)yCjs~9h4T6JFxn2Tbk40DL-*\[)b=w!,"^Yj-NIgR@^DR(tpʓNCS>3C!fH ^ wE%Kvt,@:X1>2VbGt&uw; [Fc ) `e[ $Hu"aE R3`$sP duBZ|@t`<@!Bv_@4D09N.*a]\mi 9 k a_$gÂ_u9VD)  Kro+LESԓR1!OᎨ ˅$-Nhc [籖0֡Su6O1xEuI;@=jfW/Η}p9EyVtn2_.IF$2Gۜar#ω0@:zL2)6 yƥ8f܌#?!^S6%g?~v\s~JEBk+$Hiǘ AKpf9AyMEσ]ytq%}6 n\vػ#̚`m;|}mgS量חḺM4E+GD Y&AFfJטG"hAleF͝QFG[rwD`p'3w59x+x2Mc[^ų_v(wfL 5M$t_+:.hP2\\ RY½u2ԍcmghE:xQ{н,d)zWr SQ Tr/`L1ý|:p`cAӴۭA? *|5CkK8>.IQ2m>z1dO 8CAA( L  !4(` ƨNe"Nsi3V!@! BHzB )lH?0; XH>];dhṕ"w6΂wDutdn%v 9[^M H]ͪ5tג,5h3|*nZS4dC epڊB &㔒ʊcpayꨂ){ӍMW=<_t.΅xF!ħ"{XslQ/ FOeqyKN$O _޺݋ q̒s) )s9pҊvgV8N*G빾;?ºcGCtkRE]kv{agZ8BSrX~ 0{k1F?m(QGRvUg(A4H$jS3]Uzl@d!TBgn]OTӱ?]+4K RfP1k!3HLR OӻM{vgrwlǣ)W^<=8'3Mp{ck;POy>|?{9oӻ~۳/'5o::ysu4[ =;˯9?x8x{ yDb ͉ hBAIýIխzWDec5څ[io/dMń+VG'u('+ ; nd(>Z3RQNgyZoqJ N +#21 }鰻J6/;W:T:Q:,yEI `QV 1Q`P%%91k!C]Pd Tkͅ Tx㩎F&--Xi#1\c3-F:^RZΎNZt?˧oC}!w}r8fzmALW*BsZUBH7TɨDs޹**$TF ̊[`;h9xNCo7ҷvڶ}:/#BY[iRQ \(>O?WJ$ή4hy8)=z;ᾛw½5 \BJ^5\F]^Jkn$A)JyA!f?IZӿ-~ud[Y|˷C6ws |:`|ѽ_#LSz& BS#PDqbA48,AB:&U G/3M]oEHznPK|7d ~Ҧ|Ӵޟlk;NFj2N 6VEU(jYW9?"Dh X~tWtWZ7BRiÒZJ:5J1iNG&@,:6(Eb൩184/] ! F%S69'c0l,FΚoT,>zD(74,;ւru3lsD4rJ@2>9~4 2O8ŸZNѾ#vAsyjRQ;YWɶp 4/W%MݜD<9ud@Ru^˶#o>]N7(5~xяl_Koc%%&L|ib* \(-+c">n!]i+P'Չ@ȘhƙR:8T+)ЄApEXS$cgB]Y3`5k_78S%2#ER@'DLad]5!'d23Q;"!z<9{d觋d]6յy,g咾?Kg ,5(R‡GD ^W(.֭uݎ}:*'] `B7-@R+"B6)tΊ <ՕV@g^aMzwf{<[BbmIjA)R oVF!&)rXZIdQ*N"Z8K#A9nhpI(KW@" }hLosd'T\MZJ?dR.p\*%l*Dž@XHE rX5(xƣlI˓GX~&/ xJX̉΁ܠQ ,2qZpI NL `"ywމZNRQ.l(Ix㇈yG[rHyN6;'b=k^r(?"m.Tw3b.Ms^$wC~ϑB.BS~e.cs_9mvx10vgf`| <݈t&V$t-ǒ}b [էsyO֖p12C N{ eQ1Wd~}|8`'E:*u_F}'wgӫ샷{GvA~V(S=VI'<5)DZHO`EBn6ற/Z߳]>؜AWVo-Uu ZI|AcTĈ $wa×8HO7(#묔L楡Oěb/]͑?t= F743/yJz]=uUgV8na?2u sA[ai{ק~;̝SX$iu`gj59 &XN1fk48mGKN6'*~l9i-6ڔ{#1$Ք/NTC'5tj0Y-PRԂKԴMbe|>؍ ٩޹@9x&3 x58M( sHԃ 㖅Xxax2R: F!r|6L4lc5]8gOUaQ"gpϨ+J*'&|k;?9̵]? ys=0xT zTC)Y3f[u#K6J&7}2:9aM:F7'ߒs.ǖ((E?cAی U)kZK2!ќ*_*JJHFu+08kynRIcOk;k]V␣CgTW랻 Vܜ-2f %xx'U.YGV+ JAuT4Zpu;!4](P!6{Sfl$EƺPP["cWn!x8wԡmӊ#qI0D(Q]iLPyUJ|YDX yv7˚ڈlL@\KYP\ZhhNc98Yo˶ PIcy}֣2RԸ F Q97r94Y #\:e,b؎qBނj'6jy9"2W 't:i,Q+QYw[4 U TW-CdIcΫmK %ٕ,i/j3jarH ،N'-E@4zNc-AʂA M% {UuAvSзb$褱Y?ka;*ǽ'4S]s4 :)8؇.QMK8uT&/Տfqێ8VBJAQ!] QQ&EWWn֑u* y'x)Z6NqC4KTL]d{2ϱqBY3sP ~nkn RJYw B1#EUZPuH;h/QF020vv0|o9@*Ǧo%@ҍ%2}l £Lu1g<è0m,4_nVcBA!P$JETDڇ`],Ƶ.M{0o{ JWLd:[nm[!q;*p /e,0jPF55ૈ;R͙<m&VU'QOk d_tO.N֦2X=`D߸ ]{ #AzuC.Ay1k&! 5y]%B*c\z:H Hq2 jcYۚBhv5+5#-P" 5D/R(V nhVH0cmЋF1s!DEH..5vE1pY gUUBi5eHtjDѻ舜'}{Svf$`OJ B}$`@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B $$@B}$z.g/ֆODB}$P>!HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !HH !vHO#S{w?fA]]4kͽ8_Z g귵w\i9[< \wt:>[۟|/-6kBs".Ng>C,0#z܎VNjsG?9sRo2:*3~?#~]৭]v WyPH:/!7U-iͨtgH;ZuqqEmqwigtӓዟ/~cg"(P|H\47x*1.n}aCdƚ)tΑYRv8 6pV"r^Oc&Lku7nH]-f/Еŀ2Zr\9ၥ%K4N ]k;//O&$*c4)e_f}&KI\+,)IF/yt: ؍<i1os)۵gc0-97FPb?{3-., ƪDž+P_u\RMk]An2Kȏv_>n߹)>S{,x\4<篗pvcDž#~8Ƒs[`,ohmOgjlO󩟐Rp;i|m猟 +wrzmRcVSV%f42sKU6Q5!z~9`o̕ ݶi杁#^&dnۗ6UtQlM_8>F޿87ҏË㻯x|awUeaԮ0tY)(W#_U ͺN|qw1=\u|{FQ%g9l6sg+xψ@-^B3)FUмД]By!km~@O5 kÞc79U-HvV ~>#Rk6UO>iVg?߉8777zl>uuj'w垚G^]`Tfv(wI͎t1Rn4O5E$q̓83my㼝-ܝ\Ԇ]+=rxϒ+i(`#IApdG!^8BT[)Rˑ߯j"J"%jYM$pgyAPtL hK-~CF82^yDpnZ|ݮ(_5V|e3"3 Id2įND|xL)y(PB@0S)LzSkjWlՎ2'eG k ã8SH4tKkN4=:OE(\nH9pJF۟iֻO {'Adlr)vѳ(ԫQ;!Q?ƵQFm6ZGT\\7e˹|(Ql12_EÝO=1!?Vj9d:Pq)U5$Tզ0H]eSwﺠ=vS9`RxR'K'˒QFRCF$͉S5]iP+bB~hW* [.Z2]/QA=ZH<nThC0lh0/׵BxgM[Udl#qCɯ[[s݄mP b AJJ]&E(FK(*a|{Ų(19/{V(P9oDʆHLGuC 5Uզ!{d [(Vc糖4nѻn!myL|Yu{=7w|xlʩM@-xSlwsВLSPg#g49dQ{@-yFCZm:Q v& V{mM淸ާ;F`)!zL^<8wȩ5梽b+?ʾZ凳$9q1j]#Ɂ "R)rf$e客e6-c=RVmml >-\̈⒐p~GwեaOOhp88L;8|8Zej#* м`!$^جlyuݴOdQ!e d6(j-b϶ cfCyS~Ů6-v[K:ھ{GIw!!P:eP94k&'']ӑB+-MTUbnҔPXz(чB[jbt#(ADe%6xeP̤KZ=^:,k>p^USwZ3޲ϳQk]H "3?+IZ ؋VYZlnK͂//"vũ}y5gAPY2`J> :dJ*/#uk v,s,KE%I! .Jd[\{^gvīt6$ kyeּW |Z igCm*Y2ƙ &$$6(ZQ1 gL6R/:~ ;yEIXNV˒p 8![mn8trnQ# "4*0+܉}>F0 P̡p{n"haV.IXk,:$ .x^E~i\K&r֊.!6':n}aΎ>C S θdFJ.a:8ë! };;6 MU+*(huzVn2H'wtF}  38dMhFPx VL {-g NVg AQ|I|i-FcFfHJl0b0R, [T qo S|<^tw=lpqIa׎*orݨsQ;r\Gh`$ѥ/;_!O?xej﹨Abp(ɏoӏN~?.ɛxsp9'p uSSQ3CK6Zw={kjq摒6גAZIy|A(w5O/l^5t&ϿF5? &.owfeKg7!1v@`]''&K$+EPoIm(e]j%eeA $g&*s.gbvQ$ T]_Hkq*iG3]\ \rި}\J\gW/6ѝ_M/*߄UԝIݱZ(4;nyњ+%a,֊*kl&U0dz!O$l5"G=" 8ra@zJLjaeE`K(zJUA_b8KRY|&X/Ԟ*sV2 b'|9d`"; ^UNO,Rh8WXc* ƲU!8 ! >#wM\ ^[W>ʬ|F5(iK\@BӦ)miG Ϳ}uMEmӝ'놣}{M78o8>ޕ=uf5Lg'fNwAZZEvzܳp<~*mGH__WeF{ h7!WRn5>JNftMF>H֑K3 }ӽUn"P|[/8@B\uo BxԾ=a~ҿ.SCM'3% `6C'3?-a~+zʈwH$IJ 289;6B£) +h4w_9qR3'7ђ.E.uš#73r)bK_;/\K<*2[t9SG{j.;:H\XMo҇Tӊ;t:n䫦]|H^5ˋd&+:ԹZP_+?OkK'<8xNJ&e9Pa 6`%%0&md&soޯX/_Ky+-u@:Q4 -3<#&33Nu*x?7 x;43 X] B9QD',mu?bM:[B(^@EgnNBsrkgy +ǘ=)R1%6Yfh8/2 ױұ L3Gvsok_ cf+"x:=S DѺ~r@TOEh[$ul+ QdK o_)+ѳ-UD@F|)Fcq& ̘!x#Y6^2P޵Qϵ{;l+$XH_ l1&H A(8s2)9 R,(S`C.:tj$eN6j76OUkb7% 0M[lXȒTd9Hzi76%*f^YBYv^1*(RlPFq2#P8+D4+\f)!J]B,r =>E% h)gW Ȯ!kר=w=cp!9u#ejLH-|*C֭BtӤ}τ= ϡK AńyOt'$c&Nlq)*JES:K "sq SG$A+0C<<(xD6 t*d#Yw!,S6+z@GKVemeJ#(]?Fepv[*6fQ!DPrBzj7 R+ PL067ـ9.4y7Tn\fr0gB7j'965IbF͗y,1Z~&K̳d-0,m1]񤌽/Y&2 q1J5☲"rZ#α@.jg8F䥫 qZȌA] C.#C^x`kDaLv9!=c"A&sJjqZ)GLkB<.hMĒ\T7 y`ĹȑSiK`trPp |p-zmg=zP2ŕ)9Ӡ*.cS&Vq*FRiEkk6{uӡcUy[t+i6mk?GuaƗq09ƥ%*E1P)YNb43:9 828ujIr|tx$KrmTQaւg p!}p4;u8OΡD#Eဌy֬r+0V/ >]u iV]! w_6 U/&|Ut;WGDz"lԈGːc"I0U-(uPSCOv%0 )yRX0=J(U\&R<(3%CԜ*ŚZϡ>b55ŨPygE Kml<hXOk>R[s@&zΟI]s{r jw<ѻ c"NAIX*|F!Skļ@!#Jt %X1lk*X\$lTI5[ Vkݑ?zSf\ME84Β͡9MgmgoҊ}?3Dw&(a7Q;䬣'tcM՜VZOXg%&^H OuTI e҂MMrLxi.| F)]ZPƄpS&q"IΡG'꜂TIY֜hQ Bz`f]q4jq!ɹ} =ӁՋ`} H}Bwx^LHNNсAB)税V asd%JЪ# 86WcԏU CcZ**E2p-4Y/}8FBrbBsIX  $ >PQنJC:?5t!NICXYVe@g 1%7Ol +܁梎SOBHTL(xH6m4𳁟ǑPQB| e+iJ c¥I"$s)^39d&9ygnNN1'Ajrn9jO,3vtuw}tQD["[X5N:Õ Z`AȑBI5>0k`oHN;kd4\FƭFxI.:EKTx,0U5g;Y_,i]ɽF{toNQfDXW4hku|RS1GD#N a`TĤQ=Anc̎ zΡhX#;A춣auRs A2{'lf'^I7Bo.tv=Kqt]h!Z^.+WttzQ$ō8H D*p=-:-e $ttJ=6 !bĢF04 Pfy",i|8F@Yd13"e:C`TH9_4O!Un"W&0oPb{-\<͑< BU-=F3 *՜IЈ<_!t>F!N418\6(j:j[xEdn}'BܩF- g'd Z"@bK %h۽Fx=u~jXo~ C5<+8d0Q{5@(Fք♁BU 48ŸrAcĊic DF)(999cO<ϳrϋO:?|ayƽe_߽T>a|3OT+{3|um GX~/|32gU2wՓG//[?GW3#SGhm g=[?ތ`o| 5~$ v~i)?tKj:mt* >D 43^i<Ӯo6b|oηЖ_qkM/_Lk~Q G*i!ww2̥q+Ợd|?'Xq>N\vު(m >\1ٛe {#Ǖc?XgO'*3DA&CI#tB#S)ҳZT)3ź-uΗ9\OTnPUM6r.Jm'o(G,l ʄݘ1ÛaI&zXT {EcD¢i^ь"7;Mczj6xky|³f30o <9 8Z$Gh$HG$(GgES{R-#_SM582Onf}w5bBhso3%u^heИDDϕPܛĸ9U#jz}@8*L~kU} Ɯh?Pt_1݀d5£U|n> (3L}E{8ܗkE;Uu[g5Nbı qŐ0&tء|#.r-?ȵ(B%Ny)D%嵕2dw$g%٩RkycVk&6$.':Lk\xaxptC1fooCՂu&g\M:@ Z44xNjvīc՛QQ*RZC i2vs,Zc9}lZn\{Ӂl75>\ũ`|L\:q%\H y\t:Fc$ޣ_oE 1Pa@ ϑ3CU߇*}j嘙LwX) n)Tׅ)U^^Uw>s9rӪ?w[T3AwXWߵʽౕC$nf>ُ=?k=9+HAZ  tPPBLq->(2eɴtDQmh1Y9!Y"NP$f+ ,!M'=+,'YkSI 0oK唼8 Z> LN%1EQ2lG"gfid)f̐l3G'<>}#OEۇ WsLjN*X3'e?SnVcSX V\}o<Ibp7jݛGcv֎q$S0FPkhbzʌ^:mUū?.فBvwvepo/[ 2r'r}RSva!n#Pdi*.d*e/[wN'$.j8LPʽ@Y?+tֺ{ȢǶtzގ#̉ԫhN zYgk% ;-wmRQ?i oַGkNwxf0uYk)‹c5.%ιVrb`j gZ{8_!)Ro.@5ÆQOa8Ϲ %lHBD oU::Up\ @*^u-`\SUplZ>>5 >ʛ;nO3q˚/l w͟R2r#ْ]-%u8bVC-{L|xԶAw7lWvϝA IK_6W?mwxȺ*-G!Hے76Y1 (Oset %Srq:vTH ui)6㍧\J3YG Bkshrku)je KFKޑ&XN7-|lEIڹ  mGylN*&F6%uu|[6Kcx_[1P+fY{[|'|.wFJH-tZaLfpkH[c\# c5 #B%5%g~f!m*F)v~%S3dClt)CQwi-!PԎƬ614(; hsUa}pzĨtbzD{ hw(qU6[tԠ ` %S>R\[KOCHw4ϪI2\0-yTG)Y76E5BȍK6JH!ÚݜBOZ#wݺ{5nj9oP3|1 <m_2XJPլ#K I֍OM(@ZRjkѩF9>tF A<,T1  "r VY0(!Q!sא|r-K+zn' LwڨLPJv.]HHd jS^cG3`Qn< \ 4s0 3KҢiD]v7\]E b:cHm̭l 3.tlcșПt"`HN0RPـj(E̡m*2@n70(7fȦֹ`i` X7lGdv@=P3)ِB&t6!D|d+Q ʕQY[4 ~,3lsd` .mO L+Y1wB͡6󉻍u=P?w0cOEc>! Z B@  H5jTu>:Ln-9f "+G¢ ݄b LuQ ̻Z lL6 _@k`Av-VTtg*C\`( UT h:E{Bm?P 4: y )7*b"g'v/ʺ@J*@{m!QP,f^$W%($YҠ,EHH(k 641r$FǺʰI&v"PZUw!$M'2Й1OA=z /US P,8+ .uRB|7sv9 F% {[*f=f>y;Fw E n3bja&a= ."Xd6Bt|шjcU@+0]@M`=A|E.ȋAo%JP$JDMkT2.zGpK:-|C $( WLe:[nmG᭠!ۀ8RK7 UG2YЩ֏oT>\ż;Cml:^D!w2}>X@oݷOv]D]eOTG<)}|6FD2=RS.ϷA[XT@@P))@wAK(80.adt ((7A,h&ցh[Ru BAz=r +;c_31be@o(!D xd+)`Qm RgQiiU(- AjQky\Af By( 衲|,H585JFNԬlhG 7=i"Fg4@ěPZsC{~TA],erJؔM1|P_j6ZU5oy.mP% wwqW+]ׇSߞ6mӮW'}IY 'tuYpl퉐61st໳ţbjѭ~֚cm3y(Y#8zSGF栯 Kl8,D4E'%\cۊJt0XRt*zxq`|Q7z,=q@>_uŠZ96NY!0G_&o,oV2 -].E#!HJw:HnPUTu`\SFS8&գ-|7Ɩ3(7 ZULFm:?3餳3 v(|AjC,!@?mɤ{6X:=801joyp-4.!JVPPZtÀeg΀z122~.6O 8%ƇJF$45 =mFAX2:a$4 |caќ|nV6Kc1˱Xf-$+>ER!6C$](Z͘- UXCrwV?];b BI!  T v/dkqju~el%AQ'tL_JqlsҺ}EP@IJӹOBoXQDy틝_7bc)l6YqH~N Dw8q}N '8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q}N198` g{6u>ie&NW3')O˃cYgyN?@\_G})$a]|݄/xϧ_@cޅ(~kl]-*3;7w߶/בnjOj"yX} & k\NaB!G̨9zT|4=~k1Z?܏W'ӹ{Z?i!S֫~KEMF޴wͣFz~<Jt~vz4& RQd 0y@>Oյ*Mvi&h4QQ~(7w]6QmjDS}L1gc(yt(7i=^}DFP*P8o+IaMN ?*vH<:/j|H=* ,oLZUmJTc4Џ?jc| O/~إVb䓇xozos:K6k|i5ƁS)6T__˗7z~Tٴ똨<96)x4QHӋy%xK|:tG|!&;A٧C{5,\x(7R4C_}>ԕ-D$7q ke(+B;Gű+s>nV^|uxg=eH^iN;m qͼݲE2\̤.~6~6) k=Qw'|t{o%t͛+1?^7ӎ *"O~٣q %\ǹM6~=6w[{|כ Gm_Z_5uvWߟ_ҿߐ~X?׾b{^6ƆOkdsW]]/i~QNև\}{GnC>{|{}s)\ }ꥁf§ (D3\`9 mf6Ea1p?t5I,Ơ{.Ƹ.N}zΦ144)rSM&d;(T٪8yt}Y5 'weHKħõPD>-DT,˳@[2w!}2}?gmZ%P[VkO BO!L7 {m=6r'_o}s٘lKKmKwn;A'GAb҈ݮ;[A thwbGe&sCI)vr'l1fᨤ "r<:I(GB+Fp!rKjp5XskBx w_z\bS ,ɲF!Ǝ|vV ;Ƙ qcn=O^"gNOD؎ΖKA0Y.YaF*hc7n''@iŜxZ`|dgZn?mԙI¢˩S!p͒ZXHT Tiq=-y @(&y!wI`6D@Yq 0<0-NdaCxȊDuOWN[\[Pbu5[wVt{? v 1 8qaUHBDeRx*3r E)s G>};5Cݍ7rUo.W}ݟ~xn_/#Z/"|{\Gy1}?G0ޜo~tǴBark)8:O\/b^mG<LUKM, _Tդ:;%1 _} z#sdTpV$RL_K__.>Q2;KϵEZ{#ll{$" 9ҝce)ucJMBپctdJ0kqc( Dh,' ߁Z1 B6*Z驎9oRf}׳׏)shsfO|o?@:L|juz8_mKVakC^ݼv{%kEnÝ@[֙WpcZ$N 5I1d*m#*I-R^.R)K+CDrJQ4|JVƆJ}OC>4rE9s2Drje2,*E:x-Zd6+l=^d{0qл/'afr/ӭ8_[~ك/)ӿ1&40EZ LB$˝0 ŲT*S4˔M\2{;(MzSG\r"d{`>5{YoK,oߜwWXNd1^QVQtXqjB'O"HCVaS&}J5'(~Yo,>4j͔x< \DA\d]$c[,44XX+(sdK=7eid $\ӻ#6 PˆWbc=XUSLҜqN(LGb2*}QgÀضcQx%f60{ `CXC$hS5g=b^7Ι{`KI۠3a(2*hdG g?̻u ReiaBșKYJ=/Y*p RR 2 f)EqF<5 0ZL/կ ' ۵F] QU?P)WgZ zeaBR $Cve~JR(PX lRA'qBAK;= _0hah>BVЙzT18*2i{us-*y8 zS}bglz"]Ɲmr{ݩǯuH~xm|E96뗟7,uܺhl~i h˫7B r܊@A ӵe=bp cV/P*&V>txĆ^&u`1z]T$*L^)a^i,y)y!'w)('~~*+ѺnF\:5 _o^Ffᮽ"\H$@ȕV^z[e&2+dlv{4?+#9ixNx9w͟~wծ3kߘ"n< [Ԃgr@oר6Q=n<&j-8ON"nt7z>o{dId91M* (B AZ C7i=UO eJ7yzMKυ;% "X.8WҘr2>h{^ro0-E d7RA'\7UZب"T&/;+򑃨:!8? jne= 1e<<f`*ETHCh`{+K(-u,0))Lsx&lZHܻ\Q[.ʔY4DbM+֜ \eo<|4 GѨ3<:Cb Sܘ~{>;"Nzuh.CL"뉬#$b0,Br\l: |3c]Մ=e~FI(/ֹV-~DQ%ۮ"҅XAC ؤ̘@S4e&+3TmYlXuz#u,~Yl/ۖmY|PJv2db+#5,"`24' ,IGѨ=hn| 4($렌Br:{ed%Ă˒^8h+8Nl0to M2a~ighalh^\Fs_#/y}Q'|Zr 3lrhspRU]j% t# mA}9&6Nl ,!uNW5l$ESS@spP21C7D Gp=QU")4K gi;ΦvU,STbs.8"N,# ) & I Qa\^Utbg˸jyV2HvZu,,_o&uڵ܂yx˩QTXLZB#r5mJ=AEėIX k ,uN#}DoP^f=[3~XE1[f8(/iŌ})PZw{.֐ҟP|N3NvN jp~ PK0CQ3^3z=[oP(YIw˧lu@e! kepmA$}r})lO 'vwkcV^%,l%]/-XbV uk+SyZ;6ߐNu_uϭirG:y_&Rv_{ 1/CM]̋SZ"%ٴ2k_ysE.u8#Kxu]ϕI!du9lEGL\e+4j{ceu=Sfψ3V4WM~i,׎.ɴ4:P d8˔&J/} ̹T U/ֵxX=1@Y̕ = TzD܂Rx._+x.>FVwwn~ BY˶MP<@l4J9!N^jbH,1r)ry+A%D2\(h?W ]Mm VQP3iFBk492(VPV==`E(gm:눑D@UVJ)405,9C3Yf_XQ,? (BX a꽑x TI+99%@Kiĉ>2RSΪhA"(G34$O}ԣV E&dY.>Ie>=ĺKC;5RqRG2ZI&E1l >@1#:bK r խ y=T?_=454OjIB$TljF0uiJl Ԡ2>rNoVSN]U8xޜBߖgFn b@Ռt\^DLE0~. W எ9{[W$U•}6>7g"9r:;x+S"8B ʊk!ux LF37Htm]풁Im$䃫n0_> ^ BFu$!׏t6 k3:2AF0gp4cnz /Ǚnͣ6/6jۻ"64f>jk?c럟0jhЂuֳf\+}ZPz\ E^5Ma*s0Bj.<"6_i#L}n.Y`V.p360!vݶ+7Hē$G=ɜsIkAM.15%M!4 !9Q:0GG784GV%sjƜ&Ƌ%/U;ǭ՚DLD&|UįsP;Ϥ.ޝ#9êՠ4rOٱfrY8$Z3W tʧh Dc0Ow!{ !pyv@BrkÀd [:8FdZً@<m "t I6:AI :+U\[ !EJRTF:v :NN8޹Bic"JDpi2E P 6(g4&;}jw|[̄lo:SwuQ -dP8T`n(*',^)P?曹?8,>mA8M9~[,xJ_''oOLNN3xhaN/gmLgӶ7lxw V]Gz(-~QnW{`&sE_ӯ,d!S>(~٣yV\lD7Nl axWT>ݿk<ٺ-DMuZ6!D 1ME LzӃ:7f7ϛonLU~ 24:NK(j@PzLhk=\luI8!IJ299{ -8/ţo&3vKN܆K_]\>MnMAQ7S|cv9]%dYhmd\Ny^;p5S~2R*7X~.)%hP-۪so kյׁРNEAd?a{4v9 u4ra uP@@r9r kOTV]x);*nxƒa9eC=a#QIŒHh`>Zq̸1̞Nj5Ʌ*0&W"6itr.Y%v8Os;10yrHe}T TAh7:2Lģc j#(vϜ׾z[֠lE{!/fνhФ ӏ7(B -;uՁ3M&F:ZKo$Cɼsx]W'c)j_덯ɠn3r8/޶E{D?me^ߞ Wta@}U=+OV)eB)QD@NJx4$hb+{~׹}e~RӤSCf)kмC9-E>o:x7 Zy묉S."J}8֬lyiSRQ]L%XeZwz"$҂V{GBPOj"5{sa~,좊c َ")Rɵ1!9;b=U1r_po| WXiy=ţ}r(ʣRaI@dbbm`*Qfs3DbK>%#K4xH f 6B'3Bo Cѐ׸+]gt N0-S, @p+G]AEo3 :^||_ܺ&@{Ta^HrGa/wA/廜6R^S N*ƥuƠFCd+!97{֑d ϡӏcIL03F|Y g,X=xC+u"څMbuwu9t)f4<[{Lv,Y/͇WNU|)mtZOrUW*6(%g}nedy5i [+ĵRA*dMHo9PELy[Iҕbiҳ98ぇ4[voxY~{7Gnn&GFG: zю]&@Fcd%HO!:g I^'#nO X۱Vp$tH+숴 ^ AZ7קg˫r^/fWEνu9 KR8O:X?h>&kgvlu6Kۭ^K:兵Bb˶]7++>7˗[af)^oM/ 1ghru~eL\_-wK|?gެ۳u 5{=y~Q}1+6ԋӲ_Og~W;fƫ?Mj``Py}Ȧ=yeSCVm,+-zw4*;ڣv;5x-(9ǐ}9c'Y][R~n |$k? F~YExq^!*8cFn˩^L͙#gzy`:`a|IN/-^>55m N{=0q\lr,dl[ GgZexx^g ͎X¼Ի{ޅ:_%p}Y"׼,}E^h uW0Ci41^~8`kg7ӫ?]70s?Z'>7aVۧcWz~.7?x9W7r\| C'^%ɩA}Rq{! ?r7ͺ@?B?YzHo6Ih(^|7NΆFv*}e};茖co[\x'\܊|m7e<_]/.uvɤj>[="'vz"#߮vfU寽5Gҗb蝪|nk˫M-G3땺ry0ɪO(^ug4GEʇ'iهq5?-K23#9ɅCUlȡWOPP|9j^hdQIjLlM>JeԔރluֿ|X)W_* VY;.;_~xqY˛]mǭREzKr"UזMFFa4I YLfߗJ1=$'12fE,HrFyVBZ$5'}46(A)짹tԴͥM8藹ufkMT$]js$LBySZjC>yٯ*T2#H*QkR"cK1L.-oZLmi_G΋wn|5crIJ[׌R8* HTGTI Bki{,u_z̄b 63)e%lFZ+ $eʢrQXW`#-H?KL-#CuHJtE0wvh&M;J԰Tԯ{Hq!?bާ\UFҚw-"91[_ _[w6`u0x^_\?,U6#mJ`mxY]7nR7@__56ҧ_d0)OZF%µ4f,6Y'qaj 24.T+TRTmE\g V4r?w5V4ZdRZJbhgo\i%EՂQ#yX++nJc q+CLBWJBĊHa6&.Õ .aJc^c^Ou H+h#wlՀ|ͩ B0@Ơ) e4b(4VH(,@&B#i5KdY&ysS 5%AGŞEn0L18H| DU9&:3XŸAN2MZl;Y{S4ә@HqHqƂM:$w"=7/:V*S &b ]Fkx5=)(X}Ys, R)NURhH 2BDjJc#`V6@۝@VA Xku@<[g#nVnf1/E?1k40_g!*SNBLDD߇ 3wv3lߦO'}BӶdz|gt.ئUY#xu@Kq@82(}t,d:Prt9|htVsc (Hv9Ϫ&Z0~a-Ԓ"4D iyU$ 4w,=ӅqcayOy 0}IE@V/V>H!3X`GWCƂչI,TGWgjbQI9m,9xY'O7V`yvpC$*r]_.$߰#TK^HS@ʀv04KpQ yY.z | l+ hwX3)f4 `R%2^.&LUE#)nnf= kppS:z)Qdf֘U?k>Hr9)E. `L Lyy1a·fU(g8nؔ Ȁ5ȡM!ds \A7vh-!?7 NO)r]gd*DE2T Y ))7 !0րM:6=l '3 "($+TC(MxrUpՎ9)҇R*=O((c)U2QLYl 3zLIA!Z`rOK6"\D:csjg6jsqO 8֤Y T)JK@(d$0r$jd-µNZ'}[i/3~f1TBmE1xڀi#bǍ.LJVyae3 _XU?iDF:Lr%q{N~cI}6FJǥ>t>옄C@]&0xS'L\&!K#ָ#\R?];*:DB6rR^y[z`uj)9-11uKl-9r矷>/X9\Z$cBJ却YW;V|*%*׶n<%#@Q\G[>W%@_(X@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JV ߦZ?%P7ػQrE ԭUϾ'PX *rY J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@z@phI k8%x6G^)@2xV}J $=X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@z@C1)!G d(`q^ +e%ר胕@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JG ii=N{vy/5vq`ܹrE ۧ ]u>s>K]rs^^^ u8;\O[_nw7g0۵gj^b}8_ӯo4g3`mz9af8MoapL^_-z ݹ7ed6E\ r#5p{\{]=Va]/Tm?$+ժ ers`smy^t[ڱ?k]aQM?맣48ki鞱r ḾҜbs@68^CV6+{>!qlVٴJbh^%9e@aT `EVYqw,n9gdN Z՜( ͳ L,lwL>/r*ayw:]!-:~a]pT MMҥwʰAVpj)R-S?nF":Y?sȉvzfez/1+mH_?H}]?o` <,SCVj8=,u1gBҕ$+ Q$~I}̸bœ7M_7Ҟl\ֹejrvz xƻLkGlQ&0lND4FPZ(`e&]7VV?_ƚ>39^De e7 US~̑xg3`z!7OȒpKٶ^|u>PX~l'wAh~~79?.V4lT." vcܬ.ǿƣ;+z!MyR?),i@z MaH.w>w5`ٔmK6!;"Nruh.CL"b̥l$b0l%4 HSP[WҢ/@Q:60%N9h]M/Zt{" i U2uBjE1 Wo`5O,NFa* :mxwK:`' "N?00V+ p!VʏI1h*ᝬD2JPXPFIbSgŦ$=.MQbHKOx[qi͚E0ݝ"`24' ,IGѨ=ނ@i(, ,Fn:hJd9n2af!C0'y獯4~A3]# }UvoWc9fZIfs<,,:Hs:xc&Hr6z`eܥ$^N#ZAd ?MI@,polhNLQ*Q MËEȡg mo 30F4jf<0WUMf+eqcNw@6aMAp}D?J*5OQ%* ֱ WY.e%K37a0Ǖ::07gݛ{\# Af:ӞK< L˙L' hJԑs$GSs|x!6p8%YDqE@ec9 & Xlnzu?{#{+mmq (y2H lϓ!l:Xn*UzEY9wS0iUJ#&VBYRYIr}L#~B(VߤQT?.~ `Ov.ˢ`m$"* ੺0>h6^!aE{ n(Sv@`cc "sEpr$c!r oʺ"%s`T @)n ;HLkJZnu-?O,ᮯkv]U֞]Dn%6V+fmݞg=}x~0v㶤:ozK֢k.lovm>̨O쪯5,{>VsfVˢYvw-ʕEė UYjSb~- +HҁJhC*trf`?Kt}2 ˜IqDE eu$ӞzTR{YrJ9$o2 necotٌ. ZS&F"x>,763sTLLZ hGmi:J~qi^dɻӎrG7Ma垘`9#R)GN%©Z8p nhX"QE aW ؊a ^VUom|;U '9̀ K"+dFSA"AHK-GACa(`wi oYxtTɈaw$IRXg1R$ ,TJay ͒42C{ǡߚ>ִ C[HNh¹ђS"iwڃ0q{A*+H=V$@6(#ICs#6hPDqjb̖aϏ{ 3ɻOGi0+KO~8p.|HKcI9Mhs:.F( /g{-~jX>i=oqTljx8886gh5j߈j@aPN9y甇WҏO(˓4UlQ6pVSx$y0ߋԁh'W3&VkXa8o\wEiܸ(zkω)`>6zpEҮMtvv::ؑMlTZN6M^qߦk(po7Զ,w~mT?-|Nנ&[zTRS<ΆA^];9?~oޜSf_7߽]O`c8"Gӣ_ua܈=g\Ur֜ð[Waz2DshgU]!.T8˯ 旋#a=m UޡJ۵ Q!0a@P0U|vtM=CEOvTm qLiIfZ ``r!OCD"%R$٠ [:,#\2#h`E_:$Pځ%#9'o%͆K<b[(f/yleÉm ,+͹:9Ԥi/8)?UzOyKVl{՚0"V5]!6[U,յ7o{>9g`E:' ۡO-oՖX%]VSKAјc-cS)/##!\9"5Rwvů/qNܚEr 1tNdJ<@dRo9PϜ͎5G XT'+1uCŬMjM`TYd O3$I<מp.ɀ'W4 XP4 c"𐼋jbǞq9{+^8}n3[γ>΋D^4ŌfU3shӉgQdg%DU* Q,#{g^oyϱ|޳ȝ&0o}rJ QFs+32DSM5 DFdZIci=\vp-뙵DE]R.kkJRZ=x&2O6z\,`yh7P87Z{K!5\"QKVt 휴 ҙ&YȍBȖXdpsD{E#5ͅ2QF^29 Jw̲ foox6DC1qDuR.*G^sh yt A6Hv_] ^9=sϾϷF >B t !嶙g:1:ͻ9ˡg~x'ma,Ejrn@J@V:TKi@p:)Co#;_79):c7Yi9vxQOooGܯPs-֪iT۟ t,On֙(z<b>Ĕvvf .1l؋+gqc} 3"Ԩ2QSP:Sxr~vs/g{O5{%qR CAyLAJTBAyK!%S@5%Oϝmzyᴱ9}4.fQ=ʜM}f=6^g{*-f!:'\ puP*hAF!w$#ܐv * (W,o5եr<)ZcKn'B4m f^^Iڔd~W=n^5-zEY>fi ËAYYbmɲ`O%N[4ue@a#==R}sc,؝CWYdE{7K$ԇcRpԂN*ht\FD{'l1(;yzs(GcA84kУ^/ty_d8&+ER`'AT@ާR8蘷k)xQM$N$Bi_: G z}o ӄ$IS QȣoKѪ` >P1@Cf jR bw.1xΒExێhm8T$z -oϯB$Onf8w!N7lGg ru5Šs_~e%%C['o4Y׫w{jƱie4m~aiCkGu86})-G#xx>Z: yF?}_7ׯs0%E3 8ø@ŌҪ" :*pHmɞ'UQ<+?\/26gyF_6Z$]:IxI3B[8xGw[rodbw]{MC?f3lԎ\]RE?LR;enjq6xM8[c[4zPdK-vܦ~f4kS#s*M~-S!h=frr5ޱ٠6{҆n5Wޠ)#afn&olg0 cYTS2UE6-7j ]ܙv5-- ڰtvڡf3׬w|}*!graDk -4:(<ȂٓZq T{^^SKNPʒJ+Bs"&4Z4zP`so)svN"2t\VqR}kq^aZą{/y'8۾7_'uf߷`i}w-Q4R Tä.bB@RHILB:lW[͞{(s-E%r/QFUR^[)LxGYTEBE^Ƭ u)&C$LZ{-lN)H҃RN \5PntV(Ë m1dͅ"Xms̢٣gi&)N=k81܂ l Q!KVD *uqagϻ4{:%ކ&M 79+(vEYo;Tp LWK&zq^⸄@%rnGY2y YJ𯓞80V,Z~4:+IYm iQf&9.46J@R \Įm+hJ}BH~i#m}YZl7{O۸_!%Ю0`,m櫓N#v<u˒"ɝU(Yd6A2bbWm{琮yw8<v̓qNcSIm!Jxo2NލN2TefdڼS{lߛGt W%m.<+T@-qnzd乭yYm,ntw3/˛^^|[^2< oxY8u3=w50?ż: {zurj$5E ̇7EH@(F>cC>,T1j ycl1=6a,Χ4͘W>Q'?#xX1iĽ{{mž4Y8M?H0&KRl(0o.fGnqFvŘ`^39br#'}ga)D3v0ft#2'rC+JjRpb]Q9i(L4@Sj)Ԡ{t+ŨL9Lb 0eb7o+~2tw8QYS+ e5²lCv7rb`khc`V;F&0!eĦE NI|vˍCj9ilL3*'H 2:Mnlڞ~[x`6&7Qs`uwylZaZEqXh B1}"TREE0X \ P Bhnnad5e?o d9[x=!1Q@@StD8a""hh˘RP4&)i7-:ǍPYee-=#30lČyj/'Lo.g0^v."M0:c?xıt d#Q* YF)aSMPCTbc l'*AhJ߮#{ɸ8@g}jy0?zɸ~=ʳμ ?zVD6\Y\#mgCH6Icq76&,0m^-ѧ?渕1w{q6lCl\؇=٫vs$;ucNpJ}50pfݝ=Hu>طg?gY¥y$Pl%jb~ڻ=;!Da5F?_sɊ59J ܚyVg >8_; 'yJ UXR\0G#\A~"FRsgZajM%ըK1đ<*Km( #;06l14ߡ0A ܲ=uIsCIy9[8s Uk ,DAAjJSKÀ^X Xc #@$X{3Gf,.ZV gZabMAAVK.9&c-Ogr)2p"bUi%%\Hg&ZarM3փw"vK˃"{r\)( +8@#N$TrkT4*Xw ]f.ܣ1ۤkv+DϨJ3p3k\:dHfsNdIQ 6c,cd[`.M~1̾60.|w \ YlDEezU]5Γ"/ Dm1N0SW+&Xog/lp|7!(1!'#pmc ߺهlY)wZ[+kϴN*yV?.?xYq|51+ht77 tW(GZ0L#G,8 c$?` 5,IGI :d֧w#;_À&`k?ƒnhky30? N٢y&eSOR<_s\:didqK84-lf}e[ j6hŎm6˕גgϓR\c8ꨩx<؂4Ebw^lIp |w^uuϳɫJ×x}16sgɊ5 6(=+D8{)5~N~eIYΏFSo&d4?ǰm2\@`0-L:N &Dw.W q/ۡsXj:0o=kk|`z ?2TWa@| ̈́q¯򞓵-˛ :/3oh;ax牟5_G ?DH|4U^ <>DMSK"VXs!f-jݗD=⣎Qe>qyϳq 2B{GJa;֘XSKlN4b+|dv `1;f!~1>8҄ܧ;sĔ\T Bq&K,VȄҒ9G^ǪR?kH+KcV8FВ'}2X4#uEk{40ȓւ3SOЕĽ{ :1@^+rG\yBf1(ԇ]"UJw)h%jJ b]D>\̤EWz]ńk?i6& y"r9!iEܴRNh2 ˜+xS@%aBe"NTA( (Hl$y$BKeqX)6CΚV1zvSl(rϹ-$oEЈW|j%5B,ܒbNJ%MC-5Siaԙ\nlMFGVHuL.2\oNTV4uf|vu 'ܓmdEmOtŔ~ c (f)% $T"0"0C,a eA1Ӵ}|͙Zݦƅ! oz쌃Ħw~E7#o'Jv%oĽ;;80Ь/ E&&x: Hq2Ɓ_k Θ9xÒS]Õvt\Z*NPYN'\'q(bZQ( $TQ)­2 ֤ tɫg{n|>D=9暱'Z:DvsKH^T" I|~ԭ xXSa &,$@;5"xw/k)dT.g#9,lu6z CCyN)Q,nẁmH?C7*3FnDX+l7*ڧ|$'(Q_@ W 0,2.>֔X  +zN|J iuIJ>C]~'"+WSS#{rv%yq!3>`TGoկ9Y<FCJ֙b?蹷m 95}m93۔%.53Qs<OPMρe1"H/r7>kw' g㞙MdICD#EO+6W~gܙSnp !Zw ١0 nS!ԲTaRN<))("pƊF)$Jc@H֮ 젣DdIr V1y3=80:xblZv0Lot+s5RN8\.o&x3#dĻ̴tlCEyJrYܯ2I^yɡ%gk ՙX 0ZI#0kr@|P3hq b}}#l.0(9[ɜf o1h1k6Ȁ-4F/5Я`T`OL8Jo5ڻd)&-1>0އ:l2YQp5dqMA`ZCKiƨJֵҮ^I'̌WBd8c-j EXlɷv:Jɷ.BPıfҗ?~)vD0Eix(H l }}EQRG7"dG&)Eo F[󷿾̹C)C$fEu 0+[5qceAEV~Vȡk5Xq\Z;‡"MƓ?ƣIA9Ky*m5Szx աP,xJ7CƥTRpF86&hnQ7;pv/|݈&~ (.oˆ-p1ΖO@h8_"ڳ<7>5&m8d|} =3 &hמ%7mS9o8(c[kf+NdN3"$'7o< xZfLEDw4cZVWXC8d՚H]sK5_t[bq Dm;x" YXH ,'ͼ%_%1y|kG^;cr,w9c@^E'#%Of3+/y 彂P?~="*1Eŷ]> gH'g=rm]zJ<|[0rt\ >hİAp([?Ƨ4-^ o̺(@~`ib5:Z&I;zZ)# ߧ,Q)/r =Z'+pZ r.;d>߯SZEZX0}D_8]?\nS,cB}ky*\?ZYR^e\3\sqc)GELnJ4]2dpJ/5-+P4]%."ޓ4WR>%Xؗ\9$Ů#6fyj)Ajy>@F7EƕVyL9k*V9O׹)?l'˴<v_:8[ 荎33!2= W}rnQn/.*/Zc$as3ej@x_ CL!i M;/m-1K~a, m; qtǾBo5N8SB.-cqm7Qd}L'7{f6ftnv;0nGV'4T:ԡvJa _oY"=y5٩:mejԏҬv{8˪4HS]"~Kf!Ӻ2q~l7(kkϓ=o&rJfdJA6؇J4G/$+]vGp&;~|gCp6Icf S 1B1_daRF,l:2OY[pfZ+siNOx# eEb^A`'ܬ~/u0Q*i![D]$Z{᫟pdR4,47dƒYAX; w`6yWo? ͔X4~ y6;gƏk&= 1qZ$xVpBjWcܔIƷY7`yaP{QmFJ:WN]rЏvo^l[J]OnARS (T5'?R]DN>) 뗠]ĪN)na,b2[.Vlʔh/_0Ɵ+/ECi?)EC!fUHsD-@ph?`RPLGPCmKr=-n$4`TVR Ԋ*+4[ИM=җvQٴ3tRr@PFX !8YLPA@ǝ y3B9B.*3BK"5LIJbFz:$#S]VܦD[sK3uwsU,iZO8:8L0l I]+ 8yƵဎ<(U,$wa`૖?.$ SRq& .4]oe&&cR<+K7a3)dG!tE}s BN,rƅH36o!NAp{=R(7JK{߇m)٬^x=4Eɏ%U3xu@c=6hê0'WLڠ#MyќQ+ʸ*r9Aa:ָ?ǽbf 1:8棔ϒ`mr}n{f1-;eؖDQ|5'$o_iIfNd{}t:?KZMd14 @I։R%8 !P*+[@ Ѷ< w/F*MFIq 2ThKed*L;,!/|ӷ|PՄ8At)&MxXt±ܦ"2`[*4(&Tc fiR *?'߾qcdJ(l7Y|>*IWWI4 =N->PC W0D$U@ H#E*A2~)+dVj@< -=FRZd 6HP-H1N(5PsEJr20@?Rd=qUJycFR 1:!J`l-J4H@.@,U}71ŵv <~a[ &HU0BzIE&6>ka2W'r=~R7DX 1`3Ѵa\p',q ]ϧĮmg4sܭϷ"Hz(aRW7yy{]#BYsLt wvFbd3s ıTtL>7IC;q$QDDQ'i&)PfAaYve, fRJ>޹(gJ\D`e~tQ aj۝;&9S ;(f a irQ-Д9a$8ӷ[y@$R "0 Σ d_%$?zɦ_yEk` <@/ ꏌJ??Oem!0iR޷6FDA\!wL da]"|$IQ= (&r1I$s hqLiyu>\'W:[n~2f6 "pR DS6u@e٣RUg*7C jUccQg#*ay2LTYF!J"SNzI,CHҲ_ bcGA)eRjt@$M" a3g]^Tc 9+Xgp *r˭NظV1\zM% }[SPEg ǧ|^:2XljPV>?@0PmHWZZS7hT)' R^{m#*ثY|d j&Cw0MEdG0xw{j|r#s:ouj{eҏwhWֽ;d^>0]ù,^%s;2#9k`UEo+[٥$s M6I,V(A{yE5O-G cJ3Fjj`e{1} lEm; ܄*\,",cn|^_ !ـ1}i4JtY-w:žK ~hn1GX*9X~T}1JKL~G Kcd5 } j9QguJ U)^ ]ƴN\^Q1hB|Iܸ/ zctRE}n7zY5#=@S27Be|aS|Nwvƺĥc]jC}H30ӷOu E8)$IqNl׈o&9@tZ@I}OqӨCq+r1z6%lT{sl{ SK ƫ)7 ?0-(sތU+=alj߰ a["6%M_Y4s&blE(ǑIt3l CHRH7XNA!"ZIf~eԝL noxa{jGJdH > Zb{tlƙ aq[q[s50KV9K&lU+%byß>C5uy]v;l+Q˅-{e!cN0|M~o=:P=p^qNJo]te'Rhwmr:/p۷ il~F!*؜8[ݳ E21ϹkC埐*L*ؼeY,oi!;REha|;g7o14ZJ3 3gAn@LѕT c'~z!+(e Jm29;\-BfԊtN|rcYH3,;S{&䵢ٳY\-FpkdQY^Pw aK[,b/nFšR=:u!&j `[)1Dcku7`U>NoF=FO7.Eb@(F3 &KP.Y<3 Na:M# 6omBΙho+ CVEˡ pjӳ*GJT$4hDO7v1:"9#"_!z)H!) N@'Z($\` VC\SSV&< UR";=}3g6*HP 2($2@lbE2A\} .K1IwH36o8t0X六O|Ef!d*$2?~:P&Pf:X2JE4k- fM?=45RV5J6 6%9&yY9I 3RP]R,9S ,S>./~[8rb^;,S/P}þYҴ߹sxNB-zt19K7Yomx gJv[Wo@><,jQ]wZ5 $K$~}e٦RJI޵2 ft℡6WpSE up,nnqRMSQEvjVʙ[X 40d|<:0G. RCЮBڵZښ,3I+/KѦ9#qBSrk3~X$rxY%GТ,ه@ -&4l5p-LI|طeNque_h'f|" M`"?t͞_pv\ќ"&ZFG[8JvOur@}wھ%BKDxMC*d ?3V_NOop(V!c s]+d b>}HUصŇ vTtPXm es_ft+z QW2tU(SvXxN>EYx.\Gģ_j1Fx|im&i'U%]LEMqD؏0*Z=y& D*kTaֿ#=*Cc`DobYO-,(ꪝe]byG|m[OѹnAz;):ѳǸ`O}#{|7y^ هq0>-x7p\:ʤST`oXus(#JG|$@$1^O1J;X-D}lRcаԥW1{}հ'{}ZE]) 9|,Cpx(@_S+s *ROT'Z1IG=)7G!U6!L/8.=hDNJ:,&;?qR҃.3ͺ32=12Fٕ4 SQlg˫*]HsΘ"*B^ԂFpVÌUz`rieb8\wUՍn;l Q0|=3!JZg B@u9zi=\4R͙Z/4嬔S!+71q=δ ]IkD7^GV2e6fuu&#sM 7zX5*8/൷gݯ)@Om0cmq2Q Gc~P7Y}1k*mL*@f,׉]??31$~k>䩅c~6t3z\aX?Llc{MIH0ޭLDdh+HK@`묜!C:|[-(D_a\|5X$]u,YZlz޼.> _ScZn[9yu &;<=t4z+j ;ZS&)"9S+yNe+#dE?dvj;~LanOJS=`y-uP._}{1ӂ”I' $\ֳâDg_!qs}[ E"H52 *)Ng4Os! M'*L}F*d5oB1]{vW,/=L& W {L<;3 TgiǙj9CiS(8 fԃ3ĞTwVϱ/ ݞKAޗѣB1\8aH>")w u.?3.a&Ϗb%&5S)"#fRrͅ2BW WkBhXn]P&1J$_IZ Q]= *Mbp0t^pEKaL,;-x,anmL9xÚĈ1mnw`E{e ٿZ<  ªa͞kzj~Lb@G q[U>t-qoU+MoAA!䇠!EYdž[k .48ɃhopPX`|j=j)'Bư}>AHg.Lx+d Tdy_jnǠ,g{6ڃp 8K!oY|?O]wD06ޛP幤$u.#R{ 3cTΘ9>-BxCR&d:[x!nnkiCF(.ЬY~KO_mb C *Xƒ:?g3q$sM`-UDQRf#o_j)mce -}ZR*Fp˾\.˙۰\:i'%7CxI}EF^YlwJx8X9dWJ00S΄M8[˼L/+V-ǗiQ&b Oy9Ӂym R`]V?zq&W6T!N_Lp@ЄgnU?jBti3Pt}FCdu-2Әn8[b HO8\89V~zrLlLNe{HbRO\*AwZ,zMj]E$'lwDIESڼE?:T)3>#8h&aXaztP}'zj6kjsf9p.vI~?wVKa_Da߬Y7G8 &/ϛY?@|7,{ϲc\tGo/ gkxK Ms+!t%̮\/^[TVAѵnQ!تeHie_Yeı.^'1Mz4|&:Âv|P|^D0ׇ)nbr0rnEӛ/u0`cx7.k]^iHm:BvhCpʌiJxt+G .+Ft"{De /fPYpfkA"cRb 4"'IHN{&CBȟm1u V卨e[`}{0LGHJ3Rji|l(#UDdu|O"!Ψ)L2q 0GgĞ+.#J!&)Iy>d1tsg  *B8]&i#N)sƪH8Nq1O3!ҘˈTymm(ܴ6ӷ?5z|&>,W)O aEMe1LҬlyzϘGT&roae!9c'7 ߀WvR)y*%m.. \HPg/!D8UdoUg)^NTBԤzPFPʣJZJ57BtdGsXV!cDԖ~ZJi72kcQzhEp횟* bm_ԃBTkğ"Q!cބhx%&#61*d`NT[->'s)V%NKmJeOoݰ _!cn4R$6#2r{ ߸]qs+zD(O1},ICsAam-U$FV-Tv)x 1 쀑!)*+j.?CX@Xzda@U1%ޚ{-v2IYGdx4r$(D226q6)8Y$OI#QBC:5(G{@gj܆+fWtȡꣳg¶^-_}N$( ]RɮzOwBƵbO6FbrD#IŽgr] 9,F|;Gs۫웬1p2-V6UEk]<6R!c0][Q/&;.?MpýQCpﱸNVh+6mP^Ч(iWdH&$"Rሥ!:T\s (gLJ Y)gt0r#JawM:7PRPJd^P j6qgДo=Eea<ʴ܎Varde,G"zX/dWrз,'ûvth[yJMBP2H)??vcǩe`9(%<&Mv6 -<;3>nh\&Mft˫+ {k;3~a|0eB>nµU{׊ei ӕK; ~"eV&]lej\ %Fم=-A5_U 1w7K#l,k eu9l韑D+REI"dҦ6){M@kﵛbBItxx4kT>-Vރ :6턢%O{GٿV 5\-u*:o-?c (9ǘ۸q 3 9e?V | <]`?QOΐіZM꟬MA]'9z}[C:C6+~1_:v5A1ͩa^ pm9_©]}͹XEʶa5f}(AT w);1|~ΙՎ>"\BܘVۓ2!3hyIFy>}u@J[+gj6o"/sHӕo>>7Y0SmxW;n/4{Gr~Ehk;l> =0kxާgTR26&uR%13LR#`aY'jBOW qeC,TKq:`!Do>H֢?_ёPRaS>J} %A>BLqOtJXJX1OŻH.c0ۖRRGLDeVcL'9R!#F~#f&$:LP!Y M?F:{b5-Sx®-{+Ie tx1M|@ꕇfNOv_)V=c?,ڔnK<93Փ&/zGGm 'Y!-JЄ LQV2-rob:v"&$-I Ae/ۑDV!OH)_66Z}y>dK_)M<N=/ؗ5~( ~~~_zבj=(pLC TXK E!|6y)(Vf@Ac(qN !T"uNKdԈhzPrMcVP^OAd>y!(Tv`'wh_)9_o]˕Ԛ%87<8Na*%2eشR5K> c$Q_ C4mm:t؝oK mL-g+p .>46+C|jEJ )H1!ĩ)!f|ܔ6r4| 6bk-*nd%m300M K }WۚOuݶk6( 8xibSg0X8}IGt٪V'(yݮoh5bPd$̩Ȓ /%Q6GE)|+bfGNrcY Ah^11WnM .0v)cL@^ΔIooY D]:I(uL5&V+ PoЊ* Wr)r'Ŋ ɠ^0-%LJ:IRRJ`%BP/PM‰SB*d('mJRrJ30UV@}V:Ux]jA4fSKo 9 \+\D܈<RLnO,^K5^G dYxeNLImX5&ﲤN+ ;DP:hE d^"f0Gdu~>_;58!w9I297~5퉺>YjSJiNi]9d-#x, s© CT@9%]Y#.JfmH! GUbn%]:ԼLz~;p62, )R-1'S C~ܯhx杖E)2+0Eq\:0x+҇`2W B[7r4l@P-^ eUo\coWI//)Ws佤jcɐYnZṷ= 827eфj{#!"47w{B {oKN $NoH o.äڣa1םЉпP}~\Ʃ0ycV$YfU*wxA4szNA(S%dX$+0Z% /V~rm`{,=Rؽqe\v:N!}|85eL1C֗'e/;l^'RL5}{VJgy1{ۡD +"-0bYU.S`<ĂH}Xs0Ṷ̈̌,H(BWR2ml3ɳRLPH,)zp1QD;OO9Fz矶eًA: Uu7z@fOX=|6pпR.\%գ پY1ܻB怷\Q.DVnw6II,FnZv͜7\:'w#Z˝ى'Ϡ)> *p.\c4cd䘷7e x<ϧlRT}PiZ6V,P7ʸh6eUF zKA,6Aݦh1sL=e06GcgĶhq'/:iA vx<;0\~tnP9h C:OzXhC\1WړaV[;Xr??e,p@n̏n 1v4/m0iփgT(\,z\]4^E`-&uzMxvv#NM0??p 7:eYNh 6mqsN ̿0oN . tE Mqɬ\ĩLDtX$x2{G7"8aRC,yK%N!;VggbXӺ6t0 0Sn"NؒG4ZIE[aal]ʲ jbȎmgjZo҅GESu!ﻔ,swU޽e VmàN,8֏!GbPXsq&;fnj {pn}Ol`Þ5jf`kfz,TUη.bEva70zla۱\Dڲ"i^GjOhںoH=G=pJX9 !)7 p5rl#P3/б:&` %>NLAaü " Q,4?rbɣ|8 ~\9{mW8.ȳ`)Rx|t=\!<^T&~}WO6a[-: QHQ$DȌ|^Wjg?EDpxtj孤BN=flޞҷAiLQ* FJX`+1y;4M96jX1rr˷P;VAE{P<+ ؜JeHs;5 gx)D*,7fq'p:~6ROK+ʧ?G }IrlFi|z%Y6ҦoᒘX㳿pe t:)P 咀06µh.fZh*g\Y)kwў 퐸ݟ@ hsҢ%Ɛ& Z=L L__7XNׁ; v"f܍9G/X2m|DR(,/qܸĞ韢 j%|~ixgO_ޅaDYcH\?#Ș j*vȍ.&,gTGxCE80p Iq~ L%7W]|1ޞs&M0ңd-ʉ̞:GT8;%QH-[(1^$"o?oN&˛O avY]EP3\ [a<V$fj, O%Ҡ ^K35EJqݥM :ɹo_EԄaI 'qhF(^=_qsDfSe5!NC/MiИj˷Ff jo@Hr$G#ǽڇG!c[D$jEVC}Θƃ%,?jPq_M/ Hi $4Zg)ZOB>mʹ}cMؘ*Op%&SfT kl@q eRAwd ͍i6xuGš,#OT!ΌBXqe/n`$YmaLN|U%V`Uls?.4N`SoI~_eJ ^yO'g1N~8遃 /GL q|ŮaHRʻ#NLrD82esbR8;APmY΃iO-nK\3yZ 5&KNDwpx =}rAy3(cy̠i }xsR&ўQ^_B#N,ngwD:H0q#E.GsHm K vn9s0'nA%b)8)#"fׁX87{g#)8WQtl4Q:_ͳIu,oT]̕=W }:UnĿFhK| *mK{#Xɟk +觴S-e[϶xm7Lb+l|Û;Әl/l 2M(Cm)P'WN{3î~?eB߿ZgCܜlŽ'eIª&8WkHz^1%pYZCPdۖf.~+E^gcIn7Eq`LF7*u()Er}W;8TŁFN=mڽ&+BVP׆2vm=RktDR( 2-]dX2Z=I?5H{;R449Ɵ`4.au,K`;pśvuXe3(*c#]y>eNcTreS܃:}jQOB}*7a*iuP>׃'I!V{'X3I6t{$'%>G"u񬉆}+HB:17ġꈎbȸ<'Q8hM)0/.q[ jth'>\t%>/d^焫})7G~K#[jAk9;9*z5Zˌy|N碩|v6 9u)ּ:h{R{y%4&rcLjVJDJdבI~$@/荰,3",V7gzAVNu8+*>d޽-}>TFf=[jnlje~7>uIAZ;_occ3 LuəuGGn^ƃ&1ZBYCrjSHsճ5td8B\ `(44ƙ5tz5!@pYPqG0 Ƕo.º!ր^$ak\ {]>a+.,'lE93v" ۘѝËD Y"Cl;uK^œFBst1d12(#^e/x2P0U%siaEl9>a>PRHR6%B^>j{™.<v[ưӺ+_r+g_~ ڄ1䈋8&UQY]TTWgGt:ȱ{wODs.ܡRgyPܪoDZE2^/j]]3p#tMسq.QѰP꾡^c]ֽ_o$)k R8 neq U2lK,Y ]1'&nZnnTpWZ3,ˍ̚r ֱ/QοjpwzezHî'c+F>QBPŰX3l .s*I1Ϯ3FG"X3iyX(UEf[ O-7q& veGՊ g,47s2 p1`ҷ[pOXv:( +\IdU0p!M=J+!2Yoi~CSt/WӋpÓCT0{䳋ˏ-(YC}V׫P6N48Nc,\pLĄdvj##œȤ 2C|.d6@Ea^1Z&T)aXX^n:F{[K sX )Xk=nZ͹3)4T:Za/UtSҘa#ǔ`>V'vؓ$g[98?7fƂ"Y:P8#LQp~$sr÷If װ%"L#.w$ZMj i4?8M ,> !K44CFrS |}4KծޛL(V^äk~D :=~5-D8a:7)9!s.Az.['<0m T4 1dfӥk8uEeROqmP+x WDEب6>;PgUt90|M*ef,'4sH Rj>pLy 91jJ4X]vh ;ޜ 1OA!2^_ҏm38}}CͮGXx> ZxȀFhcT#C|rmķ Fj9bnaXӧ 9#JhLNjo#]^G.}"4A(21"RS%T  $gȀ OBmsPJ] $}{ϑ׊q  x#һjEP`2^[8j2nɫaqs5"ߣʘZ,nn@޿9 FBzp5,#lI>YE yˍM1Zf`n1f ŷۘ@b5bݑQ%tE>HZ/ǡ#JܭG]I:>ں%@ujr˷{g dB+'L@rձ8?^FŮ_<' ՃwM/6ˉJ-r8e$+ޯÔb9n6%-nVl =// $ r=o~?\˳H$p h-əq9@ESc^{36_|e}/m66} E6gc5k;D ug-p2i;b.;⎸N I5ں5`蠵L#T\)mqI{Ryy*+P@ Ts"4:Z5ei/ƗB&Spz٢-.I `!0&PLbm5c$K3:&IT3ky&y7mٞv&2)cVŠyI;) N9M"L#b36xn$d5ȃ(,A)Zh$109!$/()uHy4Lř:F Ab4XRmHMi$tkޗjM$TLgcKh뼷'A,e̓|aY̎Eδ"u/յ_#;.~Zlu~ xr ƻ9,wD FL{$"Kec%Txs`=AAߛz#@r}s~s8o.V&z‡'PJ+)y#@+?_Όj4(ph=y 1hw)h6Et-cW.2(4mwYjLg)KkVor*D/ s'9bK2pKBiQ0䘴Ȁb"'F,VyM YP8jj:yێ_JKfhqGJ&R:߾/OR H5&fS5 ~$|rAK@GAik('!WNG4Ldf-S2:]7|69mCrNM]_Xdu'<[C`>WcadxU{hڳ. 1g~e~QAn,rupf()J+*侎(VsbAx`(EGG `HəPePnV ;!>? =Eo3ĠoħLW IhQA`i0W 8#O Y㱸o1.c6ξ0}\U`ھ:WulT6>!`Y GrA9QnVhtCݏf |ys;_Lfك9pqlbnxz=k'z̘CO}1_WVKbY|)VdcY¤D.?1a0Dy-}vvY*Xox熓R0rF}LCL[%  hzxvۂ':$4`Ygg;}ӻexvZUP=w!_R19IC6?+ӱ uʟ\pNݍ>iU>9$Qʦ;O;=E#w"@owe",˞ġb4)aOm*3XOZ=tބ)L_L99B#NFc#:VCgi*{T(Gx\>Mctkzja"g !#w#1֞ QF8FBO5*h:qQd9+~yA+y-NxF3#.hnc~y^IѾcNSnv#{QUZA]sbrq,y9h-g˞5X֣3nj]Z >9G4IQkq=k3ݜux._3q9`4]hVOWݵy꫊ᥧ7G3]9P$Obi͒f@r3@sՎnohf z|ssYl0R7ũϹk} N̠i3hsF'mKv7޽=@"`ぴ 4TD& YRg^Wh 岣cZ3)4tPaXw]{U޵ 7c}l EK.a>#|o(G+Žmq Wgg:OxծeA9}%gQ$.OH<{"t.7H_h:)k,{ۓezf`IXC@Q@ cR*3kBSk~cyvnx:+DSʛ!0i3PIcI w{}iեRȤ@LK, 78# i_Ɋ_>Vc&>͐]CȬKJwe S+i6{͛Wks "bP>p tn r6$Ҙ r|zS <)r3[Ư-;?{<ԙ:2ƴ V9)&w/#C]]kҼ 54|`_->#qY[냕m*T}PjyqnDCih;BSFL6 ^=<} V]#J){s}jY :V=@rkM}yDcJ7^mx3EQ-yj3t}w,EY M8v/w,y ZrM';L^WΕt,M@0 |0xE*jtZQ}XgrW{h#a ,Nc\tPW}9LdZU 7Ar#O=+#ג=6 K#= -?U'qV#Ӯ4aɠ5>-kE4qN]YW}y{)7=lPt{F`ގG~tnT0<{VY%VeXhjg2Z=GJ ΁d8|1]ǣ$k˷lGGC &ǘf^sD+|u_@;^Օ!.-_V5i[ `OGCSѺU3̦6QdceԆ0!Kl8o3-]5H*;) dt7HR$ڥRúX1aش}5&{ĕ&-è6r b6$7n)fJ`zlHml3#IO١WhKūhjV~;j80$u_ @ M"ZI-C X!W1@Yې %- hR-5Ky~6Er ,/ۤd43J1cÀ1&$YS=R]_?dHI( @˱̊ ]?-n^p 8wpLCQXk<+~-nQhx+MhJ}:J~0TlԡQ` 5xڢHau7P4 SQ{y+աZ?ui h$wf5ۻ0 Nh\PL% vL1dc,*~o}[|z{*إ"Q"?raGxwK뺨w1Y^z(/s^tHա=<Zz7" 8u8`hRcSq {ӟV7`[r?{WGO9 Rd Cp^M@૥]Y%8HZg{lImZݬ"XПLKe?3IV􃏣,}|{K] ߥD99;z5tAgw7B_]"_~Qo}GJ;u /aY6S*;i>5F/fVgOlBvJ=Ci8^Z!|H-:_~/p#;^>D_Ϙxiwju;˽5mr{돣 'oxS^wB=`kr|[˜op<ٻl'^Ej͟1',bnCu$9YyEq]cQS\ҺA%Kbx=y턾EtAec6mb jϬ1v+0's_<3ﱻ2U4Cxj浗ȾUGULs;MikI]BsGCN4| pط_|ԠTIGs!/]s!jBQn]*'Q[4Q})1VFa҂ܩaFujg80yDx!dȠٻDYh#Z\%mh2LKoCqAS*eJ0xv}Ш1#f7if`PY:&Hl;<z!O L`Fӝ5*61L[R:/"P)$,TWi.c9&[<ݴcF͎1 slFԥ{c2 lտSP/ [v=YtN3+!cj<JJJ+gʧl){sF+ԽVg gZ DnsVvvt>3)(Is'nQKmYXKr8U!$!?4uZQ* t2"2 ɬђ|<:=kŜCOyd*#Zes<#XaK(j yl$~("7׃vyG*OO2(c 4]Ѯ(g@%y~8%W?nF.dFTݮ㴂Xy낹;_O/y^o0$c8a[i0d گ&@ŠB\CE!DPȂx %C e՟cDm=j[6dիFAU# >o&b3 avH>@JbAaM%㨕Z")K$e1"thcm7`2[ke0ZfvJ! PB0`x)]xiCPˆH)&&Ri%Ø֘p`Y$ˊO/V_bU.Y(7oQ;zxuFJ?~;|4H/3#7Tc`WY1EVڪV1LJVP|\t{ .SaXKfe(UOٲnjnprܖ4`QD+ (LLgH j 5U.^?9GO?o}eYzt+k5f8]-h!MOA :͸d_!UÆ'K`my02q.l5GM Fv]{nX,Ģ_O=1~r D% PkbƈlF$rٌ)0dZG(E.{k̛ͳQKњ7 c*;v_"E$tV|R?`")#dgH-*$Q8<6k3ԅz6I(TI"amآ2TqܔQo89ı%xtqDB y2'jT2͆IRLvM.F,2e(ǜ!:,PL1Un0I~@}L M[xs3q?~wsa=$oK;'" Cм YP@Z,jtMҞRa\H%`{SZZ"F?c& /wOUڀC6PWe5Rɼ6ʘ|։#{1/AeZQK9!-xg?L^h]oF&U(L"2@+I= ٦)K&L%fZhɤK*-5稴YeLZU}4Fe'{`sk1x&k=պ ,{^7L)թP5dpֿvi𚾿Ss}~6X6ʹAֿiiiwxɼib&u]k FoҌUJd^UJIz M(&WƾddTNL2Tt'iL$Yuu s%@ Ry;%nmBtyqEmٴR9Y%Ɋ$§dT(F0)*Kbd0Hbz'Xs.8,R`,[eF[ n'㫻i 1"R_>çr|tѰVT=iͰ,4TdTM2d,F&dɳe>Ubu2EB(c&cnl>#Ͼ7]D'Mv"ٱ}Hp~^XWզ4"aI3Jki-&1H523dShW#y1JUb,& ɚMh!? _ɰ>qyq@#F֚'Ɛǭӵ`pyubu?+TWf 2"{l]Y# Vlֱ RB2.EZ *Rܪ^gė0j84FwmOTP-t11I^Ao6Yu56 ,j*Be4t57`Qm[? 2m9>zC7Z~4AJޚ) &r}9;ũ4@a'nB,u{/כ@=Jn o_{8간r“Y/_]!?\u YKaSXyZAvD`#Xy-9CJũٯPm"wedHjU)*!N'#18_c }u>aykL9B [s8Ԅd}'xgrk6b^ORM0hiA#j[.,lV(noxcI">XS"06TcT>`OzϏWݔAj5vOVTnA?Oz왈(}xӺ]-r \:RGsY{Jo]9M24o :S#셹=׬Tjoӣ&[X =*ѭ#Hq1i=ӣ;L+i}]h ,JkfN197̜u4hfh7CQsAnyc$y fh7Cϊr͈ Vgr#޽M% 2zɍ߼}8ZŸï~fU`bO 2!Wb7ԭD_|wb36&F W<^u}~}q!_t[vc3_N>=_s!) D=lLWgFΔ/v͔j;{:t迯]48k אg])15o sq?/}T F{V\[T,7(ZsY >s`%`f }ZCRPZw"W(& H2\XiQC;{߆puɹD+A' b'Nߋ('u/^~N^~g{4s!+QӠQmD*p,ӋPQ4{%:3b擵v~cvh'ZF2|蔻>KA$G/] [#{Dw&8Ks$HbY}}  -"9}D zۭ ][b6Vde(9堷 ه!Z'nMC PdjҒjn&Xugٺ-aw~GݙD[; ݟ.1s9`2;Vdh-<0ˢ7q 8h-%9< 1To5ܧe*LWn;Ѫ-ŵ'?=d:~bNzֈ)D yu7kG,@[e>9TC2'Syct׏h,yCs;OA`UH (fysnNm7}֢{.\ B"zct~s I᪝ [ʘ֟ k #G&8P$Ȝ^b#і;V_'%|MvwuWKW}2_Lts~i͘xbD0QV%U)c%c4ڕzBĊ!Ւ ӛmug;͘0X9(;kkGyg 8>dX!QnVYY.(=Se cddkc-fS>m%)=&0*=!(v~bS5wk2͐|\3K3ƒu*%,Q}C\Rr9Qa+~30;n[48cJx@ ^ QA'Hi_N/AN9to 2O]ȁgu,GX7O{u'~Gsg@nvz:h9v܀~NJI]ۍbm+UgG R!^ΈqOP4E rN{EݠeLM47"x*.a9\5Mb~`+•kذ ތy87h:FyM'GK?eۀqۀ´60 Az`V]6.x>bX..ʿ^0k.RrVfRlaA:B( l l|SB/'JPKLA_ [X"Rv?. ۋ@Je&H|JTK|Rq{;$P6 U}K7l:kC,8_kވls06~Hs[ %Z ?R䍔\D=sg )"le-ؚ`y=MAvͿCѲ]{1ǿ0~,`&%1ayjUmܵo-6s"wmqHz6b,$!$`dKINb/o%O˧uudYWUM/g[ M0&P[xsWO44W_~eA@ȉZQ!Wa4U#TQY9` 堄ݪ"Ph3^jɓ|")#QTU%5Z~zFW%^]z͘,? -hv·\VXr[,(B1'6:/%r @N |jpyu)C&0 ޷"H2*q[el!d٘K&!8Z 1S[.k2I}.VLqSeT!  W,3 a TJu^ ]rfE3'.ZlXԚ X\VG ΖQlclVEb6R t+JHPc# `p~PrGy11eHA{Ll@LaUS !D$VS9TOuFba%9U9h~(Z` 7r[1­ǼUht}~!rsXa[E',Ăh\),&QK?dE{5mQ,n~F9M@jݬG'v1GHTI7E6?X,6թPrj DH6&hRv:d]TT;5[OZБUT-D]G>ĵűyg@DOZ+aH2OR;vmdNbJ ,vD Lrݩ Vд>Ewy1&o=;]Kk[J.͓14`ϵu7f}+Wk˾9%{ GրCJkX;mݶ>Qo;L*" ,'ɻWj3 ^7+wLp%<=25*aOɕZmMe|ml~XI+k8@Ijg,Z;sBszeeGWRwxc^G PL|ĤZ#pK9Cf)sDņ<=֜6Y=^ioJ>s$*BF~ͤ7fkPq(J>$9lURtˉnLb4ɆsYU^ 8&fil95OՒ>r|<\av5`rGgqiY(6CgSQcٔExLsY`MmRXKxD*r`ٽ)Vn@M^yAf#Ig:NF5q!I֯iw?L Cx}*}gv;ҧc^L44GUF[ Ӻ-%{cmtQ+ƻRc^ٲ[&ml{XմٚSg#?ؚM.^F0`/]e]m:ݻKź_a*0~ADRRjVͫEtGß!sgZdc5ǖ)5(o/QRKƢBEumZKK?KYdL0$Q'Hh].6kWm(U8BJy1d$`5$Dhb_\S,"G HimTy@;͐s¬mZ@ s:LX֣~>,mǃ04 ƛNIf^TdA2.Xi'jF[}l~簜^v,]O򗳽O< Ҿ[o3࠽^Z||d:=P_6gMЛd#gsw1ޠ]4zw9SlG_|f)ᠽF&g_u^ &ƿ{?߈ywɕfsYRNg_xc+[jg¦rG3aa\qmN~;ܭcο97s]|ź\=}|}ط+8oxvmwWcI|Xsv\q \+I *'5faX !(/bj UV=F j^>p .g3 bj9hQQ!jw,5jæZ93 &<]b;^Y"ݹȥO'1ԪjXqrjdo{?]Gy?kRztjy;ЋK!~-8}EK&a[ߦ8*'y8x£p!.obt+iP_Ea>떢{|jFooMz:,=Z-('}EQ0=϶ +K0Hz^֬OϽiU%HGlg䵑$fStw <#JK!x?J[=ONN'IUD{!+kq;bXӚEky!ZɂNw¤<@^zSTv¤ ;2;2 PR'ݚ1~_hNN՞DZe61b I,b@zW{+Oݖ=Q4<10>$c=6zx2.N #+:Uv6Ŋ;P {{ߛeAZ XOG59aO-oҍ\yǖZeɗV {LJq8D\2Y29j#jCZhEx&~2i'羻jo<TQ$M!oux7'?ß"1 3A1A+s5rL+iM(s2dJݛ|T'-_`Ih2p.M咈|ㅲ^g74ϤP%kJ(bIL2VTJL:KC0'鈯chP~+q?of,''BWf,i)L>>vsoNj'X 7ȋcrpUYY#JA0Uy6Μfjq6!oJkN[E-EiM] ^K$tTc $5kU5&:uV8^t%bSvVigz_1˞l@Ţ >yX[k|Iffq[ܒ[vK-yd0&XUX% 8B•)$Nm:X\/E_|wSjj.VICns2lEtY9/eMhkfv+[3?\FZ3BuC40.!Z3:>~E4٥xjZvvqBeh' }Ni֚,b,4Ѕd^Pe1Ob@`HxҶю  A1Ji  V' D\ EZH]j -Kx=ZIw6)|d?&;Of%g3K4"d4$ᮖhuJ^\!!XsC. \M($9i<)s8JK8 z4&q`m4,bDӈ]8b @^ɻQF4+ tkQ-@K7~3DuvQ@aE1XqEZR]]м?l*I6x`τAJN=PN)Rʹn6"sߍ_4@+k|VD#I=AV:;m>\cJ7!h x!qSTUH-D`*dWJ %ɼV&H`Mֺ!xh9C1JJD%3l~"n>Yr]=W/I}z".XEBd,Pejȡ6ܷ&E%0 :ZcJY&X2(ɐU!DS>H!4@ki Z34˗ /2`l\#(Y좗FH i]/=}@ز9 i|~6='́+[Fn(҆P+\)LBj6j霜$Z%C 83r,+jGn_Ԭb 4t9`AieJc+79 RȂ5^`B3BYiwdp8#VJZs_쥥$(DypDT^o63^<9~tYmǍ!2͒#omZ WBCskݞ̔2T%:zRVhKF<5'G'R 4v# [)m"_VH#y܄ēA t՛ы\!˅w(GWUN?wR"ޖŷ2Q.j)-#I(17ߪB/iQ51񜛄yTДN$[6a҉}.O4(_d7FhL "rvs{f!},^?LN%ڎvT|!7SMnlp L0gjd܃k. N u:5yK *Yxr؈QD4Fn<"yDRhpDC"!dȤ4@L, i$rDfc&2M"A%â|4MvʨuIp٨H4ԘI}镽qvr},/cv,dvu 11_/"+PfvC`<$ᴣvCbWbNm6;<~ Ng#R0wƾUgl 9fÒ"-#D x"4i{mOib!4.=8n;l4` Ȯh#5*P\qw}jD3 T@F樿u$ir{) ҐM95G}sd+9 v-UN-Odϯy3eQhԦbBHk?=Pb鹄bM{ٓGKxt)@򁦴3PC5GLڊswyw'̛tɗ?bh]۫/mO FPӞ}(3V?k>.TwFS૾( mv| nFq'*)SUKZ#١w`hH슳{v|.ik/Rß!˦7` ۲ ^X,_eW;;#/Ыbh_A( An.ty$}W;>pd'ͧ#,d9Z]<WnU.6秝6jhЪz5{(0\,e/E kTWh{E({XX֣ WPjT}i_+[TsQȵzJe Bv7˟s]+{hGl4]A;V*zu8Uv=i .#v8ur PpX)tмB zpe;(Jdw|п٩>nU7j3ٽftl1me)B0zH:_=!'kZ3(4A*P3UwbnrƔ 2i L@V00e!۫ڴWd~jrO^IF W~-;뜨m^{(k-O(q)'Գ3) #5ƘNb”&Lɖ ?d)kuγ>7w;l sC'noH}Ib͠ƾ$H4=~J혲qUC_5իŵUo'cB+sjV?]<܈0!$x`.ĴP76`+Ƹi\G ؚ? nO7_xN6K^&pRb=5;U)"\c5ۧ,1޽@l{B ,'T#'o*rU'^/dž5m̾ĚM׫Vx25r[msEtD|rSM.&6B5,=3t-sq?Q,MOnYݎ YڲggCYzl%+=sE!M^k4ד/^ pvW6,{KS_is4 ;,ڲ8Ö[5\tؼ0FÊ-Vy{Uw~f3l)#OP=yR`%;)0_OضnI'@Ȋ)rHDMT1>Qi>n CŅS;2'_>>nɴ\Tڟ*#_Ap.Aц;M -:@Tvh²vzOA,î^*&]рޘ+LSPXh5S]Uao C 0-O?Kܜ4ۭnk ' MOGZ ˚<(Ns>= ;crVR‹彌bH;X4Kʠ\2Pj.@uz(idK AbQ"K@9=qeke趖 IH+OGοG{F_fxxjնb;1hX2GLAXzcUkJGq<Ƭ1/Wؘjk-_1;};^%!Tg4s@Gsb}(/MMXHT!؞xPZK5[㱫Tf_"WPv|ڽu8"%Z*i8 7(,>Rh =HNU2-=ɢ1^[WSCNTe،[9qCN13Ld#i>CM Qy6xpxz;>{W BZѼR:GsCC&/ l0UԩVtp6hM1љ/y4K- ^ܾβl:U_@uQ<76Z%ddP%m~&&+0Z$JˉL\9xsd5Q;$@:$@R bsrbiEDsvǴ$Xt\ KX5yA`ZZ;˅SYC붙lEB1p}-6ЈRhZܱt6nZiFKUgcV\MՠcbH<Yt f\+1D\~:%3e_xٙx0te/~/ V+n~싻I/y X)E2F$ׁD`_d hB'N`T:a%EFR )mKUHMKAQhtJu2HԱB³+ݙ8V/~u_?p9tayGcǛvY%=}2㇗VΘl-&'m=.ұg焉K{7 mHV{qB$x+mk &ݤZD$ˑ1o≹%7΅^aZ_< "9ѓMvf|=͢}}x9}!Y!@v-6`FU?}pVO)>ruZ"R\z[)nU`H <x(9k+RYHO B!"YnB.ţ#9j\.ِ;0Wgr (8!aQfUH/f ՘C:僄jpǿS&lm #m9oVLSIuq8cB[I9 O˭ԝVp)F b _c˞܆SU1jY !V KƨyW7N 0k8ud=$cjpye#w1zrD.Kh{5zf}ʆg3ֳZac_d8a}ʪYϯoV+%} 9u@X b F!udzdzpHH!ύZ{[P+ٷu{W3jK.{svKz$5qy8=g&5&y#܊Pɏ;F;w KzAc2Gy(k,IC;esd *,Ze3NKõ\di` LǓͶ#dss^r2b.|1'O~ [q!T"3z\ЃZ~Q\f#?U@5>l4jǺja9E!_\lӯ`tNN&.I -fx%萴-ҾxCQq=,'UbRGT)ݘ>+ׯ J6LvZ,`RN1n~'ګ s}s7J`?48\g!rlH~6I\ű+>$~thƍއWӐқI~337n0ȄTtl_ՖxxwE,H;9AiLB!!bKdU @$EBԲSA 3g8cF b**a-DJ_zğ|}K¼I~1f_EJlYsY$JXhqB;%Ko~0`Hd꼹q#py`ҝ'SJb։!:HwH)9H*8AD%p@ !(G'ٚϞ1b+ m [\k2y[MQ+4ڨL'?}%P<`u 6 8O pT? L 4г8$ߚ/N讲w(qnɛ5fc tٝ&ђVtd W'BO'3Jm 2$sM5*\7orIP  -/ ]|r#|×r}zCrpo:Mc[Y 9gΥ*JYz]5ՙ. F'wMlK;cn֫EcvUHns=̈́OYIC_KdUECCۅfx K+_Q=-mՒG8ɼX/*Szi* @2ȗxȴUa w6@^\ 8j"L0o^xr\\]g]ҺQ45Ubn{ r.#%t..,P) 2T64T7@TB|5dDV`B|-|D c\_d5ٴIq7?݌jM5Z搏dm"șn}є+q0l0:-xQM`.1L,AX_euP)!šcC {?YED,,V8ɱϖ^Q5eBr%;o55 l 穔O $Wd #h2[V &|Lr# ' !4L EM  a`1N@Q_WC)FZxo^X#pJքGp͡2[ΜϽiLɳ:)u ׆a•5BT)M]ZlF%9a/M]7-"Wbќ??6. UƆB!"TEmYV:1ǃ#}K}۩/J=ݮPǭ>[)WkuW<޻mv߽0c9*oelu R3ܟBG!U#+B Kp2UM9e0RV7a~*~Տ,,TNg` gYx/[wLΎ{fɈ (+ʧRU=?ne(ykZV/3R` c=J.د zqѮY/b*2Fm ZMY7|ʡSpj!NU~ 0D :^(fZI@mC/%ckˡSxJjhc4z̑9Ҿ H,LQdնTKbR=|}5WLξJ)ÏUp^>\Caqj}ƒaN4kQ^Kj4MVJFNN$llxCE*bn%uˉ 0E*2c61݅ngbĎђD@FLyFdzL2{Vqyp~&"㞢d%1y n:60D=>}r:uQ„r A'$ۛx ߿f׋AqУ;.>5ETe(PO&׳&TcY/;y9.ƿ_(_!c總Jcy99@fc8{!Z?]KwSL}˘&{XԬ\chJYm +%Ց0Ծ %]~cEj4<Ξe@0د7^ia0hA;P7&̘5*4Ʊ:Ǘ?hQȏQJg)J,E)(vSa,CMd]C]VMpj!mQXj+S*^4SmXbdCo4E>(qzP^dz)t^ԃoTGː{Pg?<+Fdl'垲eO@tᛣa==9O#tbNpdOgᮧT廬 F )K1Djg%,Xۀ`ֈ+Jݷ׺iRWT:ri:rT_ j_kdY#Co` \;ݩd4w[N۫;w]ApOhwesތ<=JO.`|5Pa}67.lRM~؍nCA9!z|+2(f Cu(nH{Q^Lgs]0yG܎ke5*.FWPS"/U薎fW}aŸV4ԭZ,峞|ֳzgm{ơvS 1\e_tdn.ꪬ.MzA7_ykiE;/3/+{]~3:ФH&E:4)ڇ&Fiɧ=UcѿV'($^87/>żxh`KdV :ޗ;K%RIT]Ү h\A H,D(C@#nTPReUq,-Aclz 7_}v90g]\`^^XkXСCŊs+{G>dKN^WR v쥘JűqxeYGbf/ERt1ͽvىyJzBAY12~V)SL>4Lډ:HE* [oP .S|K' 3qj'jGuuج;v&iS'f0;Qn@h+h"}#vP\@jjsLLw`WDQFh]{X~s ̰>0lQ[<̑HEEX9Z[)v]vNW{;E Y.CMejN>}R&,=!l2[9;0xf66S=vM{H3&SmvqQHYEeYsY3ej8dל\T>TѢQc].X*(Ա`!DŬM0XVLRe_$E&84C~*Ů唖nq퇋(ĪDXDPPuQ2Y_e~S-, 3 @@!BUeU7F7E-]iQ5{ SSPheډ,Kj+ӎJժrjyMU`a0^;'\RR:lBۭ-u4ԡ&ԲE(AdmYihdM FW1`*j pД ySe H5;N'RAeyi *+8 VppNLNLѮukgƂzm {P- /dn!ab]QA AVЂ5lfяgb_&7?066X=٧p#tB.9﾿ω>LS8Q%N~yNje+>-aDPvdw']]LOߙe9b|;fjzgKkyn6Q XjN}b\eNIُ~d)h^Fl1 erzUv#?|ڕ7awcn9Y̮+*lKwb\&X[р%_LWyG__2UL|4f{Wdj`Q3X0OF NT4rl?}JpmvrahP[m2јgj]'IN&v:y,zO82{X>3-avo(3y q<.3|sPix١XwlCJ:SL> U\-΃j.Sy#QZwz3%+ej׉o(=z|zv)|Q=p{]ˌ'oiu_KA6 g90="@!álznS52>Y##NQޡ|:)G> ^SntY2D Y)Rw'&@o2?\fv;S^+]Y:AdjSgFb~^V7Qۜ-dѦ2/(ˇ09|hh'QwR }"^J.cX2} 3Qn~kiҷw/|ްݗbvஆwuNfS0^Y>ɟ̹w"Z=O:ǹ?֖W>򪖡n//%M˷^0;]WcS`V$|yp789ay7^9O73c\ds^K&ذj7^;9iy()=m{QqRxRVV:fƣzy ;>/mM?Bj'GoMݕzvS6Șߵ2OS<}ft]})>U^ImKQ F\͋ \3`:sMO5 ҥzkC'Bt0R97 >ZYd5*+@(B @THp9(VQME5"%^ ;'PInDr"@fI Ōz>/u@q?]I;*en=yQC >@q3F (4=?ĒkxFlADJkWa C;%`* 0,Sta6 GBAPT6:\4\| 9X_e>-,MU1 8={)4)L-٫YE6i6:u{E} $J|LGh|}3B7?_gOei% NcvpohX"\# 13bu3!&/梔{j8P׉y0~U[Ԕ! A5x3i|SL2Tq ("`YhMG SZl"XQע-뺐|ЫCEfDv##\ 4^\RҐS~`.w(* I㑶JWMg2MO4;3``AA}n\DY1/8*{NS}7n~ܟ`%xh`"6MNZot­=]8Ň," iE1 \aJ2 *AP*VQfK%6[t={.ᠹΈݜ&e_z[,]Ŭ}Q`=_7qiDOIAˮ5;!blqP6y_=ƹ]}m/CIn[|;7(w*p킞smYm7FqQͻpf$baWB  ,T BQK S$B"P! 5SSJI)Ёg;/Pf zorO1 C%Ĩ ?Ӧ͟m AAnŪ2uHk DAb d+2=$-T*x0RcH*R_+*$J,X˙e 32:wk6nAIwa.FЃ``->n-LV+?)D(G;?\W˧Ī5wŪlِqp!ًh5s$(IiqmR%ogtyS-?iuGP7e7ɕZLqea=H*H5Z26jZt.ЎâG1?N;!"f!H R]!1̰k7m Pi)E|*"$FHPK90FqKp#r'U=E^>/&u6~;1)YArԵkOmLJt6#ԹZVz͸[隻 ,GnI=V{Ԭ(Ysk=fbBj$|o`c>WP8_q,մ]_u/vEQ:R0B0 %>!4!BW0+\arW-NԡVpXmԂkTNXs[~_İyѾmύu-ʢ:7#4TTgGOQt9fh7j ;CoAwk?um [eެUC&Ta<3ݔ tl?o?=HS"WqZeCEl]#m8}#o N@wu۳ݩsf:٧!lՒm A;9BlW=GYk_{_; X>>jە/ۓ5]>n\ ~;d"sVy{m]<$Pcy= 1 ϓNy . ()Ҁ(*!fGVLIlÙ۷gOe{C5X'3*Nuu*ڭ^6ɵ; 5U;Chw3r$!X6xx5ޭTQ~*/|Eo~TC}2{ (`P ,4<1 ֒ZXnji`7j@&Cb%@*J ki>aviolm:qdrX6t(iZjڞxy`˟ZpPׂx&@K0:OD|RX&d@㪛eRSP׾WdXԬ?h)wZZ- 2\B--ƄvWhbFpu *afNM'S{bnp_Iuk<8pGAlV{LP To@<7ق2ݦ}:1%qly}N {Q>&&ޥgʵ}ǍGڜXA TR.3%]A1v0u WYt󮵹L4C{1W|eozp)1䯄PaZ`nx_[lH>'b{5x?u00{  ᷲ>뿧B`|6\5bonzЇJ. ꎢP{5s'hRV2dTȀ7NNwK؛hO{tiO1uEa:I Ye?#{M:;u^zce8N}b{fz$Cu9(Oz-.?5ӟ3+8{ŽqD7^fmi`GbF:X06h*N5D.ZxLAYƔQXv6>+"Chb1'9RzҸvXFYӁN^$7 gF,gyYyuP۩]ޝNMw>;N>_EMd?l|=9ߙ_|}ut8V1m/m-~;7o2>]rEד¬6|RC7=xt? QO66r@b_x]&n8F_>O OޚYFwz\A| ~`,{3LEdS6 e({Gz2;\;?};:?e~q|otr7bYt>wlky^CģH_>yz۟(˃ߖ.|yM<ߪ߾`p;Cś$3Zgx#-_ ?@u8Ά?h8d/dY(/FUgF(+O*&oOiOhNL0*yb2yIד RpwdTOFJ~ȇk+|c]uzRdŘ/1tR4sȊ^/4lSV:=uq(M>]U~O>l2H1 q H)85chM\^ 0.aa[a+Eq Ӯ> "7~lL'M&[iK󲗃:G8Y-csS,reou`,226&6όf`CaLuʂ[ޣB9ߴ_?g1ĩCJV쩮NmgwjG7X906M7vc3w˗~K:8s^ _+q`9$$c΍B׼(ۙ-S6cװ#tsͦ(2&Ȟ k ~hm}ٻ6$mC8N\ $==Ғ7vU1$bA@Lqj:Bzj*;Ya<"PG0dJc儒As8~ +JZɚ3 FM5jΐ' ltS13|n.m6WSyTv}F%U$2'b)Lj6E;O_;Vl<4 V`-:O!s0G`ě?"y,K7_~4{ Ng]1d`W lǔ`@*~נ5(F 7]}:՗W^:+e./!߼CIMTځP̝2KhB0ŘPx"a5 Zԡ1/r\#ËVJ3D&)Nx5Z-S'$ĉhGcA?J3#FV2S!/]wĻ 58ji 1z|^mm=\zU{1! EǃC0npiim.yy&M͛hބ,g_Β$Z&1JH0;2Ra\9wV"aqkr&ҵUB+LRy[Lݠ`lTzdn.ϐAb/1hsKT pfZǩ GIcb:OT[-<"ujS)Xj@ Ee%ʘH,kӔ Mcsth,iR#сE%-TV*J,H}L+XKlg. /I]*aH$b },bCbے(e}LnJ5M6 S'EL -k0laHobAuH 8knm~|x3.Ԭ2nѐb'>۰O(HԨ Y8Go@1ˣQ#h|y X#/p3j}+~ )ꂷW+fC^FF? .BDDF+6e~5x:/9c\pg{=wf6jzByXǧ,[lb{sufLps\ұ䙒Y̷5ba^ltPpbY8H|d1Rd,{$ Tdn"z4<غzufGjP-[teuuQj6xHZEO}L;ׅPT̽gu;/I/19uzk^ѽ_˄7!?'>fqO7a 54ټjB1r#OI)joO@aq.溴yϓt#_D{FCϙz7%i{֤Ukvi>cQ7?F3uXYG(~Ef Ϙ&7ف\|[*f>\5!,hLu%QjŊ'B%FG{ޱ t(@Ν]$!YUAQ/+I:ĪkWQ_.b;^C_C]0~w\v)Ye9>5-Ϧ 6OL<{2= &: mN?4G@/SO_ L&.z8x}όOzl1wOpe(R.Uu8= HX3i[k΁AJFmC؄q/Wwf0CtEn(n\9xo|dH^5U5]@"iʞn RLWVYd l~,&&H Qz3>,4,x@4Ӻi@zdbH{%ex JWGnoSO;8+6 >4 9+4Qբ 㴓E, m h;k6EvyQ :6<%Ys*"mmplΘq)|~Dո([(Ó$/"Ac҄V ZU:TIZJ3:9J ɮatjUOe])ç;N=e\عlAE;L_7bsG PǝEEJI|ĈQT)S8P_Zbܱj5w/iDW,\ogoj$$lS%4!Qž>/FT^Pq Sf:HڒCu{J5"M!ԖiAW*AŋlBӡK ΐ($ ih'38֥#ŬُAF[;Wm ŊT&Zք~Yquqnw<̿2t& n-~M_yFM.͚; ]7l6?=` ;KUc5;Se;zߜ_5UNOݼe^½!YOIUn-GM1è2ʮ JŰs-&ø`P>Wp7{Rlvd!V@Uwfp__z ^TqFyp\10 ;ݱȐcbw,m3atvcU`=9K7_YUo~poo5&CceӍ.?_|2 W?{QmqL9C. B`wT1N8WZvfljMpg΢5xJKE;Gjg4Vm]$5؜P 1x0|;g}XN`٣j<=eQō՚f>B2q~gZTG>:1Ց1bnԯ]&D&~(^dkښ'0@e=~}߽0 `:ʳN+wxۧmZ\8iNbĞqj<'('g:'|n!O?y8f`[X?ɹx ,viWm_|| ej7;i PVXvhs+|kɚK6h"Fc)Cq) &y C[z-π?Aq0G^qA~Āq)0 O~H._JMY^u%/yJ NIuzBZK/)Bܥ&ޕs4ČU(cnY\H%41Il8c&5I;d-c F;JӔi 5(±IB? i9;GZ7R`/(}5Ję4ƚVq!ʟԤR,F(q9 dkU kݓB5<wr (w" re:׳4)O+\[8~=OeDd7dT Yr s2<43f*Abժm6vȝ 0nȋu劎'ݳe,PZ+ bS˗?Wh5avCqsPAؖpi׍DTS;BӝFVPI I0d]Vڗ cXTg TX|*Ŏe3aL2Y*$C-@5VN5rQ(V;KDŔvz í\oQ^`ȏO~Kﲢm (`NW6KZ,D0Yսc_í3Ͳ݌zUlE|}(ac%z ><W Ld*IGӔԘWJb[+Nc19B.F_ܬY8T ,h;{`6!-RazlebG ³無ϪŢjpQ81Z ^z0HzbD$PFEJ\l1M)' Kcؠ%8ǻAp/z֩M%-e+*c,1E1>*aÇ,NhKdH;SGcI$Trև@oau@LQ}J{jqF6l}q5~\9jDdC@<$TYNM*]1TfZvs"w $Wm-O`zoY|T @ih@GĂRoq'B!gMYZ2ESiZr iٻ8rWzjH _?p̓S0Hw)V_" )XL<^qZ2z b"0 k޹%wu^#G#j5o_fꌧs~9/V* V Cf#_VY9.x1Gj @,D_b 3ll@Md)~e󴥱5miYEGBl!A&OH:`%G sIK2[A13%ifS~j뵃([bJ.VZrDhh =IfAo`#!b'ZNp43`gl[5ۣQb5`ۈ' wi|ç?`=lۋ@Kk'dD=mwZLzM ig306zڷGV?su()jMPǫ?Mk/c=@m6>=m#lgӏy8+Q85p Ny}YSS!+\3gbYb}'2R;sl7-+?,9r6+3^=\gkp-sRk m?qjh# ?yf|W S8{_sd:6ozTӧUu vj > ۧeLճ։6Qφe q@YD3?DuH=w?9_21 L2˒[A;aӫ/οlu%T:puU{n"KZԕhU;!-hE (IF)IL"Ѻ$e uE#7uCp$1 =7~݃1"DMh:aCЦI*)(ɬjIeJॉ|#8$FY1"zV[QBfԬKE< 9/d a\h3gYu~wRv2Mer -c,0  !|Ds,w" !VҜj ZRd А Nx~k6d,BKLtjA4^m^,q⹊m*\EjXX7ORzQe5.)pYM˾)ϕWkybCYn/EA^c+{->ыkڪ2]f>ץ뛇ly5l^lplN/MsK#3`iK*԰J*#Neg:k~W~{zZի؞Yw _Z3TrM! `n5w$'Bo3c4?Dc{Qڱiȸ)h|I;aR9#8F\tm΢M!^Q|/RLQDۖ".d˿^mF@ I@]@Z.~~xoqrΐ//T-w:0'Rlѝv!h_k:Q:mI`o]Ea EDPVr"eQ A 5'kpaNgT8I!dLE6%V%;]猉΅Lȓ("ȔYǖ6{Zc)NV]ok\{eg+EI['Fdlk%Ro/ͫj^]V4}{%Qs8”C̺'&ٲX̙E6‰`|scdM7Z$1vrVkZB'VĢ 9iZ7 Դ$#?oιLK 8XvS*v߳%tםƣ^0\- 6l9Ml1I- FQ˳aϛFp`l'p[uM/dlT,a2%k՝Q?uQx,_g8lɠ;ҰuVӼl{6sل$ڨ{qC v$~D'2+RQ^ BN\t(,S$>Z!cPu^U>&tN_1;,^\!}7ԜH kF3N-ٴ@Ipš{c#p@dHx/`x pb rz>[VApRuEfrg,AY?Y (|m)MQZبI!$*o^ Yi1GJJЫ#Ʈr/PGiDY&Y$]J1Y4d2sYG𾋴֪"BP4G[Z}&;Fʠ(ӟBu|5:a>XrB;^o$ d+/ɞR!HO[ibŨImbם+%wJ%ޯȫ,cq,%H[ Yy%u0ZECQ쥓dJ/*1|.}sgq4#~YQcXD*4^X6perEcfS%z’e#$8Dm^Mİ1u4FlN)[7쫞瓴n8 Mz:kqkzR VR m߂sBt|Lh wIeaʚ=Z fm,D[6YE hN [ ?}$4H"2ؽMN[XՈSbS@"LrJƁN;Fԛ:n 5O" h'mYVcxɄ! OXZ %P oQgs%&!~.s YF#PG C* fgN>$ cZ&Fz Vj%gN)"k49X-ZgJfQFi0hmK3P:Uɰ\#s _L+X͘Q*[kyْdI\Y0GrJ0 B?PhoUlC?Lšats`k[徭*VHj? ^ؠdK͉ Kud 4,Q=NCyw>i#ݦsKp-%%"u WxtٕpȲ"Hl `<km!Yޖj1JRK4k;ayya Vyw$ciWw8;y֞ uay"cұGxa7G5XIڨ:i};A(=Kl%Թc"[V* ){dx山[*r|sn*}l^{,375Уs*tf#P邭H ?/SrH ٵ7/U$1h,2U%n!I O=~f,yjYɛd֖ N|>!#R"wSL=bHFW/' 4vqttes LlPI#aZz&UlykT6`ZֽF"LUȌ ^ sh"W^nErұXSNI84M6Mc-)!p2nnM P1N[evB4SRޙBÿ-)SIu@Br`BW~f ޘ#[ ~:lL9Ryލ8N|uQ;z?l]4rb6ȷj)>zD̻}?OX9\7O6h_A%e]SW -LhOlH ˚>9맿դ ZOj I"ىt}!@ѻu՗ y?3|>/طpj P5-wLՖ뙪mK (ǛXA'!`ۗfXyϛ}DI |?1@+c^S 㑮׈vt{ca43Gfݾ_"7+4}E+fG k Z Vmi|[ᅂVoJ4< &ʨo$`/@B)};EQ!\BsxT<9a ln<|n Ecu '906'(@kVHz] XΊT͊ڗ2 = .xM}j6@6Oo n[Vd0@NƜ5'[ƌYd {a4L[қM(U/--V)YohTc$ݍ\-ڰj3Fd|^CIQ־+b12rFĒ=90HDnEϬ|u#@"&SA Y"B1ۈJkI3TM~u`BL<_W=u/Qv , ٚSN*z%1G`NO T.jY~y$Z7wQAP ;8=xU}ixpػ8ndWiEo '$UX(#9"}YjEb{f%)FlZͪ"Y to7+ Dc2mږn[&8&bts&ؘ񋹙 m?8Kc-[S1! f0Aq9]THRy_~'MLJ꧓g]Զ$ 7pд]dW*>1 DP7Az߾zMAr=}F}~R.?w#E1:9{U#/Zx+|>%w߿Q^|J>X{H90٬C%Нja֝ljfcn_SYLy /V0ҁ'%tGWiGgwgP,σ+bBt&_^%/?3 o0oQ\?m0mԺW'ߤ57u8hzS`߷q@; ~>]󹠋ͧ_Uq{1U$q&K:I Ւz5j~SClj‡4&C\'UνsQ[ 꺦1(Ne8 3+2A|nyB)c,6aA"uxځꏄ"ߘ CX?ܪqJw@`(`N Z=">T` v҄Jkӄ}H P #h![tvqkyvz[xH*=;8Mp>왪Xf-W =UPs u޻ S9h"nog" PNO(c@5 #(e HHQ\TǪQPm7CW|1Q창\L9eooјĘ,QIǀn%cmmwdCE%r-crS; ˙ΘI=l/8 {SXٷ=UA ?M̧}^V!&hģ) Nj`8AIte \]F˾>G1vT ⇳}9+L u ה/݇bmw_GB7/saָU pDSZ=z)t x=<\&ӈk =%T΁H Wd@(BH^y}_a@bxbJ ?6~6iQ^u,o]+&Gxl_ϒ|Iy&}4C>MO?ěl8A]F*pwH ,.XFH37[L^?^[,e Q@޷ =Ўրn<> AsdWPe[ qK~Rm 8r)` 凥zNPd9&ˣwv,n<%/WN~5=$֢'7c0sa/+%oy݋=TF_[ WF/lXbyr^NaɗIymNCL2n8z/T@NXFq)NG00 >9(*fr &EX-+Zqi@/VFB.XiX+ رN6X.r`,΂ζWs f:;q1 㧳<(leVQUFO5*~_/,뼻츛owz>04y( 3E~gu3ِ16HlH]R ^3C5''܇0({\ ZmEڼNk8ܤ@nM$~Y9r:)lkރFSbZA4cTtkQXeqڝǛ}@KzR՚~]NHDwX$ZVKeMi=@BPi-$<}50GA덉_|%32m\;:eFrhdY2z ;|LK-ʄC3ĜzxFuOw{ԣފ$?oý\ķb9Av#B'j)ײt|׋ձC}wtv| ǯhsz-)c5ފ'Gq3!z+Ai=^`2lrQ\%`)t6gsY|X-: j~qtyY4ېOR3gXp&wFjLBc`9p<Ç/PHx 2x/h !(LyoBr #aLp(}L9dkd(0\23B}syzس 9@>rN( u 6iNtһb"x% i$7G֨d! PD 64mM"fKDwV2ЎuGASGu(:,ub>5]J,e"]NHor"YД(1r=#: (yYu]uTȵ 4W:,26%`\{k !lҐrFL!n -+z|=0h m.C3'd}dzxq"ێ|~&<&̲B{AMhA{nm!۲{`Ⱥe/DXt0!' @\%qMЂʈPf.8Ei"c:{wFj<PJ1DW!B+gE0DƤ&}ljA#&n0fD3 jFrȽ"7'n{ rQbg0㔛Ȃ`2GMB1TYc8EmsNH9xCn4S:5 9[^9whoTo})6^-.ݲFWͯM\?Rݴj~^FҺ1g^yj?J,NIzKrCzow/>Oin=zg=wNueT@UFh顮k >V+թuۣ(a8ufԺ !\EtsFwuCn2(Q1X="D[5m y*SPSp : 3` ݃V Je,*g&q`r ܗ+B=]*I=Z,.IƠY޳VX$H $1DCx_%(eMW6R ù"~6w6-:zL1N LL)'IϤFdˁ>aRB1)%z9Sp!eHZb-2l Q(9*F LiޗgIzh%4|*_˿|zr71T,D윈UA穵ג>"}tJlH5tI_ӸBK5 TE9 4JM¤t6f7*b)v4z#2 Pv\;ՂS!v:ə:$oQ<'AyȅIHo;oA+jU(I/Q\`3i T@9(au5uԚkz <ԑcO8@}xB`ކ[Zu7V)1zm]ȘN[ !\EktJsuƚuHºʠDujc?fL;[5m y*ZS&ܓ[7!XZN}@c[jFKUtwB越>FwUr,w+KԇX-%5'1bY!,Z h> ۢݰy<5K^ y s'5RY\rOz}jGԭ2@`h1UAkGD>–8- Pf']NwmmW!e;9{⊓˦T b"Zro忟Ɛ"7C`.RY" 5@["4A専a6`#KRR3eZ8+(0rr2q}0T`X 4! 6+ *ܬ\% p)c IkU63Ňb裝|Li)Xۑ{#i\PU6!p9a"kˊj+3.x0-/cIyQ+A 0:L8y;4&j.t 2ExB&pOqz!_5ߎӫ]j-4=pQDgCEb~$uBLC SÚ26ލm7.A[yn 'ֆ6k(A[W*裼V w7S${sc9p?a&3TG@ۏ51pCMk"?XXly{]?8`Z 'hC5}vER{j.lv\P}myx? ::CIdZ9io7E4Bw)DOO,֌"}bjpW5gs=+$ r j|cÎ#> ċK`ݳF7o0VX$χ'l;ߍ=%UҭهǛE/ FK;jYKqY1;[昊@55oCc)H. |O02WG$쀤ɓT)gPaO`*)6OKeښЌش&0Wנ}yn YR~ψEaj&0%'$ijMBT}(vmo"9nq):!Rd򨶢Kx@y*D{na!"jMRmlôV]ƙ h!t6>;sMFşz6LvdiahfWG;Y! N3_Ol L۵S=]""*!Ùj/Ggt+K{ک NĺKlC._0$E H/ C:1h ~D@̈99TQ{c eޙ @짇u@5vِ؄E&%91Dgߨ IN $5}(N*;C?@(,:o664Fi7K"'| *{aq1U+&X@%R&1*rc|L Jkffoj_A|ȺK0' ٝ^>nz親xLF(X[ PE\ HYҜLq*@FϞ%>_o}2GZzAlG| l~m|v6sΏu0=m\]`b``~˼*b0Dd,WXg)}^k0R5NJK |9#.Q$}6*bTW(_g&c78'P e.3, {iv`y)!ã*Ԟ䜑5:Wb`ƃ~XT&hhC *B.1!VcS+P.PǍCEAv2bûfsfTDI x8v bMz  2ޘ)^PMR 0Q쀽܃5}0` ǫOܜB {o /ƂeD* w1"sV\uxL7W>dyΙӬ`N#[\HL(rIWNhN-s;I(szߕ^c}p߁[3qCX**o6/3o; !<FXYd\2[j'QFkYfcH`Iys85rDZ(+$IGeSd\MN"_2GSg96=ܗT)TҀX:Z6ggҒ,#H'>(sy$2‰"tF@k.S̀"ࠋHdQJOeXi5 #e'-iUک,cjǕRh|;خElf%`6I(P+6l(> uLOu0Nh0&{s(PXI<ݹ 5x[cXv *muͨD C fo+ODD%xZiҳՀmbL*{zCX̂9 >Y )NR )F/Og3*wg(#!!2 .g +|z8أrP;o]-@!0M´SEJ(6O1FC쓙 HyiFCr>(>aJx6D5lD'q1CegDoIYiRzD#[\Qc>%/Vô"pvL,7Uy@Q8.G-0tg`E27?*^qVRA  wJ r,ڜz [QPG 7U(輙Ku/ Q'60jd' M-/E}@6K$+6S2O\}J4x0+0+Wjpo{s(+z  'zS" HVShd=L{xP,Qd=<>c4{=5!_UH~|DWrH~|›DW H~|֙%$|]JhAgSeU:~yZ1E SmkeÙgX9HZ&Z3ǀ&sE)Pqgj C+—ȴX 1uA XaaTe9@ueA6:KEaPtIDk_[`X>A^U\Z!`?:3y3~jMpvXp5pï.߁g{ ߎGhR=BiQk8 TRU-co/T ¼TA $"G ӲAkic'`&z6*+^mZO;LCaO. 喝E z>{ρE>#&5IF`JVR 8^Zi:-m4IVFRVmrVsKڒI4<>a:ս/)5NguƊt"_k_Y=5DžFFJwP!_tȘጙ'EP3f0XdIIFsWyD:K=#^`r5gs 6]fhbc0 t`4u@LԂ@#c𣯐DZ ΂ө(,T*PFhD;~r os`x|'~M^4;5Sj5}E1DK|ʊu97ZǏZjIQG;G#RMdYoԙEaX PdVĐ\(fAFP&NY5źQjŎ %UBO:GrXɃ'^Mnp=k;~n]mw+^oTAI}W1zJUk 0ra'n8 L :Myrˊਗ~4ͥX5VP.4E4JTm&E/[,BD':ڭ;0U=MhvK!!/\DdR>ɖvA Dtbݺ3jI[4-p).zw}n<>h<[4-pjp_8 [,BD':ڭ}kTִ[4-pcZ< nZ^hHNd[[lhB[ y"!S4POKR;3QbM*N)1#2TnK{/hMt[&Rb1Ǐf<k_ l:,(H(l Cg2 JBv-' Xb0q0FW3̠T a=>j4xS]C7lFW cN OE]?0-}.aA,){?4NOnO󧿛;>{Cu{ps< 5a'  n4I<_|t޻̫>sg,]W]=EOC [M?~`|C 3oJO|p1h@ZZ;= |A>~H8W84(7&&\\ gt!ͤ"YH!Ifj%!0> (/(^8Dk_y5 zT?NfWU+Lg O*ADDF39feb(aR"./'n6o+|I<̔>YQG~q|t]_һS?js߆ :e;*m60҂n{NglڱǬJz7`-Sԃ%,IHMᯄx#%RlmB/ ΦߔBʳ^ɹ# >ohk~uɆ;Bg9Gsit>vfxʜ<"!ovA4BwP:r6bNb$>6Fi ܙAՠHe6.;G3:KĔ(t"Z-(V QQ6Dh& h)>15gDf0p%orN֥5*Ϳɪt QW)G+$P-ڜEubr3HqFJHB!3R %8"g9\R]%p@*ى>+LDC`cex|N|/~2[.oqIœsmrHZ_\>,V >p dq8 bt{GBSVDSqys^=eFLmmiDaعlL /UQaL: t.x(YWS`)a&2*31I5[\ռ{xV3R`f4"De d)޼Pj̦}(Q23x%3덟ԠyJk+:xpSqe9ИY#J$+<BoGJGJfF/fj.=k}( THX1M\u 4xmG۫ZmdץvSpFiLtRkQ_rF6l>5AxZ.`ݍ˛VR8$2dBbʿ</5~& w.1 N,v[vZap7MuA%yR8*8,dĘ-j gz"ZQ 6O?o+}m4oDH7ă90njv(~yedӴūiOsҰWr;uw&G$Ibj"`B(7>[ T{yG]u%c1s;:y+UL,N8qi˜$>q KF]d#~5fiq=$l+Tq+ ຬu_S 49޾-NٖȖa*riƨjƧkOf?4)zt}ZUx=U*(t {u1ʷ>JW-.Ր>#v,Kdǣ>dߚij<~TFB+<~~:4?qͮ;cj<]d@H5k5C>tOGG!1'"3:8uŌk!wP d7"Ǽq/?x?xc*}8MH}r*-"EX?m><7[r݅y0ր|ޜ$S47hx3;?9} u䢷|k;RoU˾=C[y:޾lߦd_[:|/\9 y7=yǯ(3 $uƸw&ޛt9pmY 0ӝ6K<>:w;Qi Nyچxu]oL񊠣]2/%+Wel{;}{ή}:>1UWi.h9>kڝk0z ݔ_$z!1Xލ&Ϯ&::9}׹ufu}Ӛ3Xn:N11F')8Utm(]%bjHPܡ`kA8qe@Nc֓wPNS-[Ztם¡igniHw}bu̝EH;Ë! 9OѣzGcΕ%z\YXTܦq.s `IԹPN 06yC(,r?Cn MEb1'$ 97MC_{zYwv<|FRTۣA}5(Y(̷cPj=u68Gl̞SKG~)/ *LX^ؼ!CZfc?emyB4H=˔"312s'O@0$6[R"෤XI4G01YV'8 s8!AA(QV" Kt2%$>üI=yBeʦ&Xx~dm~A L9\s?pVH JC7%j>6U@ / dR]ѵbuJrMDcIL,З5yC$>hx=Ԝ9iNhL}`ύ2Єj]XW}{+>%/6WssE+sspUT~M?':էL5Ay#Rr4NyAPs{>Y! /!88Nnp781\/4:C%=a@Wn顴EoiLuWwi5[Yuy|3g9O.mH\ 63I6ךkGB~"Z)VQ E 5td];Dj\h`qŒX P0I̵΃8y&;Wkt;{Fe/f{3WMQ;fmkoTgڳ8BfgXy- !نֳ b@^G1BޕJCiqUtMLYFڊFΰdB11jbLp4zS \CL^{-8ap/:l[wE$6唉 je—$Uc!L8sg#2"cNR޸Q%[!'T2in6 SԨ$Iw:)AL:L[o8Xm3;5^6|q" f lL9ӄ$eBka*m fX{j~C)z`A1%w8gdCg_] ߯//"^_4w eE!p!! ٚ7#d+#Οv XɦB(,RC%Ci:\1DDX<8Aoν׳cQwT闄cڍOJp'V:v0Xwko!Sg,fTs]O/W $:nK;<Vj6LqT}NK V,W!> ݁q$9| 3UX٦5ɉ  HJrR>To!7:՜\Jm-_c iqlx<.ͫnK)K^@|ZͬNdZn"U"09ZKq2[Ҭ[5V 3ڑw3iO: wlDJG-).UNCC4% eem}YfLƳJ&Q=u~{`k=v>YGg h!+cPZPC߉NJ5W؝A&T#gϞ8j9U ־<| t#1cYڊmb."آ_۞.M・67 O/fXu=r%  !ԝ@]ú"eξ]rC nLmGȃlH/w5+8mQ 깠[(o.)a.5"[^n2J[5͎Kqm86!_!ƻ~rc㹅_}& 1ұ=<N0 U@yV)k \yA% gI6+,~y5*}HYTeWlg?T8)X HqH@`]N, =}L_Ӎ("l0ºQ=$٤Tlk+n #!̺^7ZpҶlBZV$O/_Kmc^8@ndbF?Yf ,S4KOG@?|shB0O0{-V&<lI]ɞ{ϒ[L#^q2kCB 5'[kI}H-vHYQ#A)9&_[` &5“\cI} 1RXΥյҚ@͈B}?/f:"jKY\rmP^T)RF>wHni-)q._ ܰ ;;qm~Uf+%BحJ4҉EG^ ~ٌf4I0 tmJ(Ud)ۻ 27McNe'i}2 -pZ8N3c0Be~&8ܓonX9*,"W%aKǗ^Uf9{PnV^=zZ09c!&b VTl]:k>Vt!w%W24:FT~3[e@ g+/1t,?)ݞ l8# KbAol }"-XP2 Q@ caR epHvᤈcF@,e8O[u HNyϞtGRiRKʰJvu׽f{@*+sHq$5qpxL8f#tv8VdΎ:\c@GT7vq1<&XaD CG$n4NGP, 9bEwdžb̅4@pcNѽym*h9o*A_{Xq r>o}9jފS/!uZWx:1rV,u?xƦFip:;1-Jwq^(x7ѠRط,"h]ēGOC4kkh =_rJm>V5E|54()s;;Gg#]d΅T2߹1r%-`XrExEp䘻 t})؆8dZ.x6%g}xd9-vG J)sƒgru9N}d6ڞ!Ys @Ȫ /tP!~GG+}@͸Wm @J#Z Q%mvH"A7QwPcʠrTa>Ma3@!JrMʘ5{`ߛҩ7ǘB`[8I̞&,lb rAeet:3lKR K垓!P&R9dbOZ@4jsnLI(J]4%PF {V9!rdC@IuH}~ H.t1~w1?@X@yD}49;QYR?%$e$ <\<4~)t9Ps<ؓ&c"8?c@^[T^{ႄ qb_E:ߌ]]}h.̯;P 2qK;JTz"6T:N ̾裗YJ,D,w@?E{vDyRь.c?|YU4?T{)9t :޻iT={^WU[?1nzϦ|gN3 "%_{{2OD՛=j+\*dx~~Y?#c(>j3Wg7} &w>~x%i6_wH67})w_;JX3ߙMYp6)0.R6_o9i/{_ExKGZg 1P-x"7ЊW5 bOt~mWoZ]ko!۫"ROTRW;I! r4޶đ'ayufɷ{HR:-VF@UcDfZSWzȋF㏷7} Dzb#81B%LHB7_LNj˩WUzI|7Dž& [M?޼^mO6qOft 0RzljG|xSI 8plG Z(q WT|Ȟ/(EýV3!EIy<âx% @m>vDiMϑ`M#eU tSr0%%1_R)̍?Tul6ۼ6<\|ŵ8hv``xn< Ubm#7S9}Sngp`#ghw_3&MVR[-i4SWt+cъA䆏2Se Gr]8_Aq6be #Ӑ` 0!Jh:LX QcLÙd_v6`ȱdE8^6 y /5OQC9z*~RQqy)JH1})~ N݅x:)BTMTy5Z^~4yz~В6-q[+y,.^[&~q~^:r5(\"p!aXz2|/Tثr_|ȡVU;lQ$wyo6U~EϫYerᯖb[- 1RROUU\hUX&ZWK0eRmC]=VJvLYOnMsx͔ewU[{Ϧ{uz+%vUtӹ:}/%k֤yк]Ւoyݾ+b!xp'huC{׫%E,2n_Z ɋY4uIPF,! MmgQm{CW'-0X;dZ9ʡ>MV!NPӿQv4t@K5]z vO, %d ;( J..hVڒD:#A.1GA$Ɠg̛z`}(ڠrgRhqtٻ"d2/cܹ9&vF=)mdkִҺ `N.eVIiƈm0'v)xTΘCy!]fBϾ!mȂrs]$hM= /;tChtG5s%idDK2ZqƜI%xհ4gYK[3Y?ڧ'^R7H!]GK.ūs_3:cwbúxVsljb@>3E쪳0Maۙz^ 0$q;-V7WP}$\s?*\x"9 XWxxʼP֟ņ t,(歐`=!.]O zG,#Rcx}-Q9/.LX_םzU`?+<ƍ4ݠ8r|oh} ] ̋aX-Q][zlvͧՍ(mN8vv2k>T@dxR]3Sty0ŕ3}Ģ'q/~TVi_^GJ.e`S;Fi7Nwz@S/!(>&!Ldd","8D@-@"@}Nkum>+IfS틮{sױGQ XfZr$v}gnKQ뗟Ibõ<ǃsb3ܒO|w?'f5q7xۿ+Jf[=;ߖ-ʻg5n#j.j#\毆}&ǟӁq,aŐO䇂1GI"q|j%~"p'{hA`OGC 6pA>Ylr9?+ظ _|53fBTL a,d3BC!Bd!GiO(6P8&\G,&x8>5\xS+I6d7m-4gW"v7w,{CzMxn>^s"hLi}!#>qDڝC{ԛ^_]!Xmd F$$!,Ih2&\꧄ 04_ .n>Vr1n4Zl\}[1$վ^E`-ZM\`P\x o(94}{{_śi@w\8r=_썰6DI(\㜌a_HeqUl^]PEWx溺;FuAZ-gec˩Um62Q~K J3B/r)AeYmCg/RZ/xYdبjt/)|I^|3f~Yxb35qgg Zf^%}A oAߺo,/IUWM%+6?Cr:/4zmScԥ["iG[36,hO{G:2Tv^{ @)]g޻љ{{Mu\WrNJ*pM& ZXG$VFy/ _1~.:j_#Žkb$>TR~b: N;D0z\ H INPQ,"Q$<% 3/XH2&ɎxQxB)(c_ ,a؏HILqLHct0 1B#Z,uW5x1a|e2.< I~۩_YzalmkwY@2j`2^$%B#_& cJ|Bg$Op#P C1 8MVm6^x_ǑIFd50yy0v4guJ>˔TJYb̔,hAe2 8:%ƈ,$25\%Df\pwݞΜIIb{Q#˛:+&W2;<3ȯf1.[t&#.6 +ŕbI9zsmt|\qWQ#䯾o+m })|;|P{ΐl#n?/Kuv͇ؤ(FTJC@glYL{['A(Yҁg~HO?و]y; ZC;vn),$??d<ߦD4o]484^X|pu2FBd,G2ef!ٯ?7%G^-Z~{FV;7<⷟ևϗbCB_=}PXO Nr!6 hZU4ý4=xina,7 :w_0:#TP1ra 6a@5ſ"}W qߟdNxlBLf 2X#ID5XP"pYE" pυ^(ܣܼ5Ia|lLeڤO0kDdlV&~E a:RѦ?>"TY J+\mΙ^L2U~`gRE2q7J+JF{* ~MD*K|L^V_ܼw#"w4YĶ=}~jcF˫~,a2za낼m֏*v'ނQ-Hm@0X4hm)KBɫWx|X!~ hXDy[[Jm5=DCYR5w[&*ZI(j z(ozCF%GO*xACp~;&$S0Hv;I$uHQ+FoWEH_d)d۲ta(ÎCHQ8Qɍr095RW!-sdWip˼Jg=c̛ dT/8q>+aꪭ^)_kcB(R4Ќf5j=zfyuJy?׌P3$tB 8&폱T'cb{=v0z81qW/̉+Ҷ#?t^&m$m6 !A+{ ~- 1{vDw˞{qd[g2J5btc5M+8"{V3̿s>(>0y%g A&a~vnf?}3cT牓8N,9ڄm`q>[[/]/8OܤZJ牞";Y[a ᩲyǙ=ѵIO~9%^9 rr >ʔTce^ƃAQ{b$IŦ &S3Pz:8(x)=uqL[bUn:\BV ź 6Dk>NXKbPFv9A$u @t9"4%'3qrEJހs^d_O|3#Gfzr4\+lƨDzAO#PNO$B?Rũ1t /8 bƂzK`.N D,'%dOS$qk]bLdjo[,Y~fg=+EۣC d0 %woPV@6N~rͥx(!Bc/m>>}0/ G)* j8hxAԷ|QC+Rh1ʬ6tRi@#ҁFw%E}N IBM*}@rrhzOr>8x1 #sDF]QۨID ۾> CY XŒ8nQg0t_bYش|L2ZGYk ^HM1Za% !@d\+aU+޲L..rәy9κ7PpMwK*Unn<\݆nAJ’M-KԔy&KHBJ=ΚHʋc(bu"R t:2.H{1c!U=?vJVS˵fCݗMkTpK0vbpklӱɎtَ}dĆcI ED`u kbX:H6l,NmXJ89?nMC5ıQ.ri#'¨Zǵ%('ذۨ\W: N1 Ej-Dj9Q$[^%kXq&)Q`O5R$/x&yREpZogSlh=Y,t>Akb嵷U<YdprLΦ%wg[ӨKn }-7ў;jr6_lˑU1Kv֢ۤnnØUݤ$ q*;|4Nƀs~4ةm;F_bʈc$O/]/_<~ oU'5eO~|UJ|ly k_yBƆ+ùkmUb^)d/2z[HV[ƦkȔw$H cJlz99RHF}ͿuXp%b"%A1Ұȃ$L.d&1 4$z\2O7A9Zъ $Eq" X*%&Zs/RjĤ?%8d"m{7md"D(q.q_#AG5Q! Zf ȝaeH-7Q&A"%t z laǯ@jtIM/ kIqA1&HN&Ɉң@T3@mP6TNbl}r˜< WSlKi:k[yn@9}o)}/#V|yAӫt}|/~+oMqF60:5Fφ. _1kp=(=T[L%޲.ki.A Kk?|־==ZiCt&Ft9zu{GAI>[و.GOn :QbzCG^ͦJk9T)80HyRoqA,NkI(0ź}g X+Px`.j 8Y^EŔǴ_t&*d %N,FdJ$H4;! BAVujL_|8ͽ:O AGQE`Cen2*E<xNV@L$OW5#"q u Jz)#ElJ6-1[Ŝ~׭b:yK$PO*#'%uks>amB2N0Kv{OQ9myoO֊WѫwWY2Sl2E>˳,劲Z;R0>UzԸ1y9s& RWI,GdNr 9*k8&{&}9O,tf˘5n LL#AN: s*>Za5CZ %NKpVcѾQAz!Ż}bnYC΀bdj6V0e@#_@B҂v%Z PiF|5sHܖ@ng."y:ŗk2/heX4kE8,$j$-*4CpEoD'T0T4ruI'#îMd:UNkѓvOZq``5#UksݒBp5U! < }yD粍X|CG ek!QYl'QŌ-(ّ)A/BiP!zd Ut:z4u/wmmR*=lg+uv]q(Ufuxql忟ƐFHCb32J쐜5F{BPD[x.Lv*ؒ)T\<mRlKfxRXm02lhNZLg 3?pD:'2pA =% K}nglkፕ]lPZMKa$7p#劤|ڋUoY f8J\Lgk8L:s;0%ik”Z>; 9&!Mqj7`Q,D)YaVrz$E4K輗t]R-UD'][{oT.4ݲ ꐐ#,*YA]5׵.*ryPEtrEL8v&vCB\D3d v˾ 3j%#j>蝳&19(sm7emwe)}oa7"CMiCD5sNմLs,f/E*rG`9J*#,68ZmQ҃]P2bg8yC  y,h׊0R{uPDYs]5߮\MWf&*2ɕGSlUՕYfɔb_״[vAvkѰ`0ilBj:$E4KŹd)w-UD'][&5;v&vCB\DsdJa{}Qk5Aj)| mG k@ ~0%"rtx XWK.Ud=X)".pKQss1ǭ0 d^@MA IsHJx00 ^#bƨ5w^]5_' syPi/N1Bۮfg]AB\Ddui]n ىLT c%TMh$E4G0R8׵QC.*]'co-"p7r | 9r͒)l,GQQct0d NYHcEO,+2"+NBXkQytN{&H,bF#ʐqV|$ L{BZ;xqT"jP K72Z'ИIOb2p)6 i>?,Lgt^Z\̈TNFD뙇FkAR ljZw)]uK ?JlsW$|kt@_j)E^?ۛ0hR8 nF:, nh{ɇ&M8)`xl +BDx= &ozͧ*{)͎ ͍}:l&FCjt$[)=˫eYK&j&z*-AtGv Z!ӛj%:om~R6$)!7j˙pӖ> XI7f*sKJAj#(GB>Ygi_4^W2L75O^4U%8*ŵa?x9M\ ySj2cĕ.̷JAu~Y!@cR-y9Pix?{ɧih8ܒit34ק0_pu2%tg-z6{߽w9ozS] ޘf2O{choTf!`9=eo`qh`G 7yZ}?^g.n"xoI^[aV\yZ%Ύr'r'eɲ7ޠH{U;"@RQw3mQD͇SY$|贷L[ILoIo%qgw"0Ő,A(.-2R!E h13 #KU:eL0< .t1c|6r)؇,(Ͷu>?Vu زW 19p-]LFL';(tmq:ҤAL|4ZhP0+񋯯̗"i 8g\2(w_AɹO`2¾7`PKEv |Νb@l`X5λ0 #h`o}ds yg;4Mp6O$(jFE@gi,Sb tfd$M?l}.ʮJO fUg?yC1E흫VwTjK_\h)کz"0D%r+5sA CdTs#sFKHe =S݅)w+3G8'WaA{I6pw/S aa~sY&I8Q)Z@>J{RrY%C }JQ=A !?Ü1#CT Ӵ;Ǹ 8!] :8l0]9?jh)rk! Gz7L4+Žy?g*Ž q߻hVv0HlC,f2RG]_j0@@P?Q`{Ȳu%5TPYDKxU}ըԖ>8A>4&Ò$>ywMj5+~ 'Ջ %HwTG]qTt Q)hnڰ&СP4PHiNwW Q0)f*\'J ͫ#Lr@7 E6?BKeN Zg5B(-ÒCesGDqkc}7Ƨ^@b(bBAY/SM| )ĆV{CQ_􌢽ψTL*Qr[gEr^$gEr^;J2Y&:m$\IacZ'%#!ւ)%+Q(Rn{K L-T$&6ٛ$S`W&"=Ao dx\$"eã*D=OLތlK?+gaMp Jiۊg=׫?~4^/~BtEhOPX7ۨhbøy- *iD3L_1H6jZ [>Ǡ]=?_{ZcRa|׻ۖ Ӛ 8}s{G F>-x ~^ lDߢ}liۛI)㊗^]}ˑ8 .L+m$IEÆ;#_fz{ f~ZWusG"eꙁ%frO%ř?SJhz6 (Ѧ+~Lp Uh;KJƿdFs-ԎSSZ.qJ'\eScauH.A>qu/ "Sǻ?OB q{?}y UX5A{k 5;ҖS!hAw,)O =4Ϋ0/ƧL-sH!~u5s ?_suYGTF uۺKjYhTn4)%=*V6⏟tU|d%FJ͎ՅO]FSL2UNnVd9]]˫)rdQ=Z!t7 C apWNv8 EK {ǘ\Op |6W8N;m`Hh̄B.V?/y\?'Y ݯ}>@~wKF+ՠ$X "h!pS-dY-ߺw=CUGY{sjy,L=yəF!A%UnԈnmҾFiR;#5m BSR}K2q[:2)'8% ny4vF~LFoh LUZeFVGui$3J U8 %5Vs \`:}BY+U=\A`-YRJp"p\!xiL"jMݕHA*(P)qRT6P$Iwj[VOP͏n{4&ͿT[(z 7èAߑٻjThf4rdAH Ϲj^)23O@t󊇴0Y?u*W(sIkU*6:]$?vR&lC!ntD+Whhn:_BMKN9'M `T db#B><%2R}|H#ȗV>6,X"`H{9y02:oaUm<SB9nwK9+/>.>0 Vf% PJ1ZN#Q|5|HR5~g(%e~#. DծQk:Sn0;@wj"wsa0dp,XFrDj0*T<6"J!s\Uh'π!Snކ*̀噗.1jIJTS* )}~B]4og,n7 .*ӛ9_"YDkY<,>[>0R}$kA0~ IweUmA ͷ Rz ᤔR8* Af׸)N@}z:..9 H=_i3f3CU'|يOVSۑ-Z-Rw6.: J7)RcTwGtwh0|wRTRX b \xI#ش=fulޢշ ]]Fz~lΓ ZC0#T o(hEy3Cj ֫8m9YC$KWO= >n~5Ҿ<.~2Si2i&t4WUo~xd4TRJ]rWZR) MцpćeF MuJ;)1"&ITqiPޖ$G)% !`TpP^k\v>+Zn]p /7·\i*`n烘x^k\z8N7(HUAUdZd#pB/q:\h2:Izmr7_kJ2IOZsׂ9IjGM/波@&:J6g7jB0yLH0ʘ{:0,JHa .}=g'ufWٻl{N%{_+{<U&  xR1 3x?u73Nmȭ'[wMV73WK%b-0|],)#hԓ*nﶟxX-m)-ƿԬ2bzQ?M"l ۃ%WՔB0O.J\!Tӱ,JIG#H;8²>E- ) 7)C+|O +܏dHS߶1dVbP;A%Fu e P ()/߱L%uT4#\ FM+iF,mwfuY_)KaN"StAh{Gőw7^Y[y9ƭ].Y⻌ΓQ:{.d#@ZCЌI0qDWа&x{O G's  lL̫ ZKɝSG. anhAEb\z`zHHCH#dik., ?Ӥ>UrUuoZ.b5*ǯѠ1#ngAI\Tnil QSsG&6trWovScΏ[_~?>҅ܽMs- }s\zؽi}/׌/K\iʱdefZ\B pVjVgR?Jm ՔP>1 .I4E.ķ%C^H&hvT ~Fv{r^"H0}H߈2Du{a 1RN(qt^ 9SpJRTuv/kXb{hL-JIv)T=> g0,Ci6@IM/>kDᦛN hհuHB+B* ɓE[!j;@t3W2 bhG_D D ע$^),Im ˒J`ZG#i%nJUjfa(Q%C,OaSƵSAi3Ew7NLþ  ,6`EH=`V󖜛P9FZ1?/Zg=[$FBAfWuK,Mύ0 Bf kk0|VYܯ&zkau֙H]iu&Z[B@ -bi%Zu9#ɼ#M`tq­$B9„X\ y @X3v'm7F6l5AZR]:-,R Z48Ub\ ץs-('ie+)u{;3Elq<B4Ri1 K%a,Xi`Jt|>q'_ANNZJLS^BIŢ3.Dxav 9w"U[ V:ݣp_*#| с- Q8aNV"[ńq2mI 0:JwLy^ x23ę NLIK͸tRFi<ͩt0*}iA3LY@ PGAQ E˲4x]3ʟsuZ0bl='95z95AVN;uu~-:E'Z.vڠk;L{ >Ņs MjEӊ}@bK fF@/"Ɣ ZXi?ח!߰\΃5ăbP,, d Z9ِQ20#xW!mVRQʙ,W1N }1aJ sJ犮m9B8țUءï{]dWf0Xm2 MZˤvIw' u(^ }˞W8 ,wt˭W ++Qf%]e{mN j[Vc9CIn9CiMޛ`4:[$OWEObN%ɷ"JC:`SyBn qmkg/ :/T$fk`?׎|6{Yto濎k[a7;+ǮGfTw UL^M, gSϫ{ >[^">7g 룚sޕf~EZeQnݽj<;|/ ~0 1X`!.NC\$^jXG<0=V}]w>>DD/rN;W62VFJ*3ԇ?l>kv+Ls+7H\!Yz|Wozk)Edۣ[@6QKo(hFACP+^z ]WWߺ3[6lM=J  4Zc =7M8]# 9'.ތ8=ۑ}6z%ktn_n/d`nDNS0oGl)@D%SNaaO#N=}_N^;v jIEiV>t?ޛt ȳt>U:ڳez+^CwFGy{gd<[TZs&?)|>42b vy ۓ^7Ru33Zpşsz0CCWm!!z 7:ٚ?b:,z~e&TKдÄl?Q 3I ;39E*-PV+r ϞI?mI%r wem]w%λ 5rF 0#YۊJvt=նe ˽uӝgd] l fFn$ܼ l,O.YWLmӘ MtUAQ C+ j )@c\u-dbi=o+r~sr!ti[ƲbNO7:lyVb恝ZòsgN;GF ok΁P$S7yhZ7Vm'rҖ, XtĂ? DT'RʖG7'H3T{H]ĶMEtF|1%/@,gӏf~ o@#侽CՇM\ӚY>'0\dR;YX?E3U.6K' *1qq27~zyqd3`+*mYö,7_!T`t`Rn9\TY/DYWC0w* H!dGEs|('}AyL(%cJ 'AJ-Fz7XFk =C M +wKC QW r'4`ԖPI m m3ɄDe@q*0Ay@ 5JHHZ&$R2m5xhf%)8C J+ag6mApo daaAİa & ?F"RrCi1f5j6`r.`|AXsl=iKb*(cg750i |Yˈgs0SogVNtt_6?)~XV)MVi$-GPl^-30]LhRW78o/\F', #mMPXQ"Kd `)ʜ0;=Pg_%%w R9b] j`)D^V N__=c"j9p'{I\x%iW4Z>bba➐\bq>ao+#4ݵRA'̾P($HRXT@(f%%sS'X0C g \uswV8~WZEDswQOwP Ͻ_;D26Zpk\j6w{ X1 N8*y "NAqpL ;]R&DɰHd}zi廏/<6188>Qq_qqB_箎/ZCa88w/s88w)/ƚ_\0ܱńa8Zpz~15&/&/.W$bh]Bc0WeI NHE< ^|GjxbJŔGS/fG3/f _ NLc p}^][?[J18z lƠ|A%UV.k$L0F\r}R>&2a b>^G+|S,~\]+<ػaxvoF>d4Y/^}T}K,9eAD:G]p Y~ݱJUͨ566l T*9ߝja'RB1U[~○Q]]8IE[;sȹ%|ëVMfvbG;TZ=VT'oN5V֩S';b3q\kv6nN5|vw!e:9`ߙjO_lR-uwS$ iC1.'5Nk/sί\c6ѭYl]T ^/e zk)c]yqWo=֍zhR1h":clcgׯ/Ko!04j Ƞ` YR0IjQh Hp螁!+%9o4x /-)58+ I2 <80cN ~:%b+ ;^g6| gy0fX[>gM"\NDx|~rz{ia>]΍=T[ͦW6>tӮ.,'.JߨtW79SiTKo? 7j,[GL ]Wg##:obK|\kOח䜍S<~ww.byN5l )\ |kpXo&8YjpѨd{bM%EjS~XCa'L>mN8a qXjp3$MveqY,7 3ݖP?H|@Vwni,J`$[P}I5j>`[V{ga "=@b ^63oFͨތj~W쿘п9{Ueʲ hQm*`=-lgU\in_5q=~s_׎h_ ܴRYs~#{~?;s5(\6ɴSsb>q 1ќ+`R~K\/arrYBX?o™ia*-sRVSq04|*fp g0X;y(/[P/{]]ǵW作\%>_o>:]|6[9qeuJwDoh] ѯweIzآ>h,Yc !O"%T^o$IIEXe :J""S+^)Ur\OY1n;Z`'.,&_o8Z=}S :!)e!‰T:T(tB$QȘ,v㇅Ix7MZFwnb+ÆVt~Ĩuƴg&2g&ǯ v懼]'S4LApb[2n8 %_+Wy?l042w]Â- `) c'5;{ܳ Y01sbvDom cMNmEq"7WC: >}Gpl=V~W-Zs:[5*lXRUfbW)<,uZ$BYW6\n6ò))w9~׾]'q{Q)(ŖWZrD8ӫC .?$  Q܈xNEy))4Q=Cqf tFT9BiJ3$305ez,iQ6d@brtM_kzJѓkJ2:, rKXXTM!W t:6b4 Œ_7a|5)j ČڸP\rN"}Inapu3>s\ׯNO1C+%:r~ُ`O,M}x3K*{]ȣ;bnFߦ' xXβǠ(qX"ba &/,kEH;eqP\: 5mhR9/,s`B{d] Y~ >p>ͧTC3AD"΍b y(DE! H=)B[ CHɒw 2_S7׹f:ص;W :X'd.vrk3Y( ufqwoB[K8@<-s4 Nsxwi.qƩfTgp9f9e eZ*K%2ng"c~|yxWՏ%(~:X -FpD%+1AN{*qHb:#~>qџ Eixv,%4AF ]?|5EAXH[Dj(_iX\N o,”+46"!AUއC8\N5v?y|.iIȜ~x4lV|iBu6࣐{+UT\P-3-ZH^6L"fIO!VD)16 +28A; c1¿tsذ<;GCux>h&rX+@b!1"-wkZGFjl&즉DWLtE%uJ$ET3z,5B3 |%Q(|ZY!IMV|ל?&2}Ї[kɂv3mp\{\LˏI#+=Z:(JBbv45䤫3.7Sk9jj=eZq=M5ZѾ6oIe M$SJbHzSEݵepOR4Q>kȧ@hQ= jUGW>Zb\GGBq_47TrGT`#sbJHQGr'&XÚI<t5舌y0"5yI̮lN"q̙zw>UBFdIF~]+L5PbL㖹%_4"l9SI)JGUHÅ!*hM0DhRa9յ"–ᧄ6k h뺪7 (G|LZ'Ԗ2=匂R.#~嵏qX;Ϝi:#$f`S!XduVz͕1@t-(Aj,u-fmo 4 2Fv!m蕄i@EbJ &xAH̃R\xbAtdG $S/fM·e46pk*( 43L5B9l֞#H@ʶhj󼎰 i_јRQ" N>QH@)u`‚j!n@XD FH(#+ECx1Xmo [@·kZ:ұY%vD;gsϓfKB,eQġQ=a&ԋ__4"ag}APQCQƧ>^7+1V|5V?D[`zXe694E_a˹us5(ep41 ӫʦF} Sϋ/ƲSng\r 4TýS+:n@٨ ,-F؅sٛY͜?97MI^6fgK•8]eag6 jqeju{}EC-I6]$ e&H 3y'X;%Ayx|XuMz9 7c@#)N!9Nc+n|=Δtm+cvls8`z(7׃ˈ_ !1FR:WJ*S9 Ӧr"oG5@.7Js)UȜ[)PpnKԻC#MSm<$Ne$mNYһ'"{M,-7z!1odp =Mkur^"9hJ oujr@߼ ŗ*,Є&z2/2pNWyvsxG5fr>iM fl*k2١l?ɓ_?~yl hyxO=U jdb>_&Wr8oYXt,vL.><.Ygp>6Q^^{9Oi%O樂yߝQD;"#R= e=h%2ΙfP \q8S$*A*/nѥc\rZrСd }ByYjr}5xSQ&%=ѠUaEV+\̮PknVˋ 9VF_;[$d'Ok(ۗ/bl񰌋l^dnj#dYi*J"_;~ WgسF+Az `T)s*YkGsj9,f2gC@ f6%szblƓ&=ܭ @-Xr'msr?*y^+Kc*տW~1T.KXot>"OR\J#%* Eމ⿐Ur~7lq:E_zO fJƁ7 #)>>,ЗNpޝ#Ŵ?hk*y4BFLf"^.U5v*?&]118=u9w@")˯>KXdj5m67S)mA$#O޳OY,گOh0nU$-:j#Q4v%c<#$ 4^*ȍhܘpL:mZoQ [ܒr؆4ykxF.W\{Ѐ9ύ#RPvr hFT5 '@K8a6-FBOB@!w_$I/DIs-y(e'm+oVhcƉ2! Q$7ET|EjR9'le\ٌ@H0ԿFJ8^CbgR}tM!|-l.i%} _fZ.;|Ԓgы25ƥFeMւ{eJz4ajsQpggUp boqw[tJꉻWVq@} hNw?GWg0UFI1],WOP^R;q{si4=. MM"}:(=)3$SbEO_:7\lNRݠ ,=ЂpJkоTާeJLۯkN@ $ t)bf[i,!`0-=:PBp]GHR{I4LpR[e睆:T\4_/wvcL_]Kcq~ansq|*?&Lʃx鮼GgSۧ=ˋ!LՉBT\EVRceP1e8`zo#Ck"=_x=mgVS#=16-Lp&3#ʤʹRB |+hЇ=\!B/?NMazElCdO@A42F?4R퐥0F[<cܫVd§=lFӷn:T { oH!*I4gG3B|^Ve9e t/m%/W!FB?;=8\Nޫ=(5& 3)w67x)2A:Hyk{fH9' G6! Xj,iݹ_kBِ\OK S(NlHFńRF{DŽzZQI g}`roqjQ97}$F jS~5Ih^*V:9ģyWM9>2V<˸&NdGrGq9Cr3$1Cr3$ IfZ{CAJZX\xy,>%{%%K)28VAuqVZc k0ReYFEn¥G3}ʦS3H8: 2VYIP^h"9 JqHRU?RNMi$9WR.q55_,h2 $TX"ϥّ(M(=!+f=e #Cafx>p}`ȍ Vy.ޤ\^u(W {/Vѻe+64}Lk iG2J, gg26GiN$d#)?*].k&}+ՖZa-f_[ʴ(`CnG|NjAY,ŏJog-7ˉځwXWoPLɄHNHQU&%VdKMYܗ7'R&ډ0i>8LS&SF> `Րeݣ]1rg|iј*Þ`vx: Cpu{/> X8a,.B^-kh'a9x8 7w!OMGܗŗ`> h-\' fY4P3Dmc{ƉFm vJWەul=i{N}Xմ`潡O݃GOɃi?_!S@9UoL 5D1@C(8&}B05fxO9\c&C̘0lt 9Px1`:Q22c!Dј\il-h6i~Trp̑4`"akHXhy?X6I= 7c*{(4.`qC֪ilׂ@j&4.^[E }57Fsfo6Ϙo!dZO^v#]Q}[x|SȨ?rOUIkF+WM'Kd,5N,FWEVL%( 35S*›B?( Z ssNʕ<1gZN\H!3@dڹfO5^_i c|n<ӎBaFV}ti~]3٘GN(!]vM~jE2UR}dL^۪姻 _7++\b>_rYisFOpNCe&#MC$Jbr7xDQ WIosӪH#nA stck ~snŞw:A1 Z'ȑGŌ#_ƏPKRNsIELxG BxL3-H,0e9B >31UKJMA Yt6):[#7Y{&?̼U@M]6PljEF#ƯX6 K` aKdi_Y_+2+`VnMnV@;a4B3x l4#AylT @`qɉ:V)(GhViTӿHgq x+I4w|Z͚նƫƁY[G.pmIܸ}Jغo#xv=] O`/l]tiFpٽ,F[e*(s=˷n`ߒfr"i3=l#n%ݣ܅7ÛOߜn:c{ l){(k|:ѩ[?Ѻe7B۫uwN貕)&H4}e`vg.^`wv׍n`FH0lBۣ̐`ձWnِ3wzC 0H@.l{s Lְ~qNoytLqzdD6;;_<(/6bҺ} rrmEzcV#Xơt|nxpV8>o/^= %c]oJo2'4G˛)9{h8vS,縙Waa`.RW9ףo?N^2B.VӖ)ܱ$j'XCfD! u`f'XՖ2;Fֲp/oI޴ef fn3i>MMmfM˿$nKBb,`&$ܳn ’/^#й*۫/}~|IG@VtӬ{1ݷpG OO޶cgm'z]uLce0Ejo4#6 IӶ׭7>06f%oOs7nN[?nȇ>aխ?߾r2l___sd+~# ~~NU뻐Nn̉ =餀JR)y䡂~wyLoC:8~H>Y7~&:ai5ο/#^F 3:*z' GAW߷#N7ߜ<.drH#^z:|!_?WˇXG _MCYDB!4=D?ݔ/) g=pn# j˜ro|I uGO;it}I\<]ğvw |xsXT# ՔK8}oFB/8wbn;a /ρW9bƑKƤ7$ "'D qhfB-9&LV1"al"f\J3pbQRU=We{uLKUf_\ hZ \L+m`c44r0b"@6b5głZ A 2Br-cFU0ps}m[aZ#`^pdQđ8 %" @ȈJɘP1gQ;0dI0< 8➙LX*ܢ)arX 0ʅ uMt/K%&0 +"p5O1 I 1bAQPa 1q!FcG =5Hjm%r*֝d6ҹOysl}}#yJl<݈YRM{ Lh #rtxEL&6o]ouv"@/"ƿE̎0M 0J3Fտ$M) Rx 6Kryp`R!Gjz"VJރ <Ñ&FkɌq1']6&P PG ɦ3(y8coW?~71' >}z|>Y[ KEft T4Mθԗ_2]7A I ̰O 7҂pJ59 6` Q&$rH1 RV7vn7%T>xӚ`X5GE3a}(jLL0%FQxMsBT!4/I`R )׶IRa\51VR&Ʉ{R.0MZ7imڪ@rm\Co|qluTXTQi-칚Gë/@DŽ=5W*xf(T RrE.Xp ^g+^@q9XXK6av3̝  _Z`sL@ݚzY՚ɶbH="`T-*ບyn`EhnV~;D[FJbwfTVTQMF-lo"B"b:#) i5!B#戊)8V1CnpI"kOqDyx! +.}8_h_bg%H#4A2=eF-5X>EɿJAq BMmqcaXc~Ȭv:((B+d#4#ZN6oEl =3_}+OSQoњU&>V~r GHn`> 7+sblKj"Kx7G?QP#e=?_J3鴭 B\ T%[^=zyyW'#aP^9θP vK'Bi4DZ$21uķx$fTB_OV݋&_OV8K 1GcP Qݾgs6F.7iԎ*OhVS_laE;U1)uI #T,-ܙ&V ;тv(jX`%"Z3C۹wT͸lM"D.übm R6&@yJ'ɤW¨lZ5R NKYQ{X#fF7 ,9;!%2"JзVE湷ls0L߮A!Ӄ=3)A -eOnZt v.NH:,\n~IG7Nc&F=t~I:J7NqA.$0ȲlqEO -`J/< $d M{ص+I_$$OǽBڴ9vaDq_&ƔXFssBt[1#(ZofՖR(%+jJB$h7^;V qDSmhۈpUZ%m]8NXoE[RJj/ldhBU =rR.5ek. !/[۬EX-lY}>sHu6[fҒOm| %vSmcԕ"ḙXK]wmO iޘ8mIsL6~24TC:i5U!m##.F +ҡ-5AKZGB H!r-`Umf_) +1Ԇ&kx.?VLz[n>^VkVDW@ЄiGLLpW-DXpϷ¤I)ⅻjehΚ#xɮZz%cR$(*Y DmVMݚ<)l[|2]o_lޅ6FX*r!D N/c>0*|i%")N3^nqlLE5"| O>R44RJވ?F7F|DҾYC*ƴe2ֱ:i"0FtF a %aؘ.WuJ%NBzh皘0dnS݋fόG&Cс@F$`"Yn. Xd^EE%6ן:19թ8jEEmWXCBK ںr $B(t|/*JHE0J<>HDtFN6d6s"VSAT*6UFKGш!9>hHi++>L(, lO BHPǔp˥X!1N%9Ds*!ա88f%Ȱ}#;":UpI 45?6Oni4e3h@S*(4d*#jHbM|O+z0nJIO&Xa*)*כN4jts]6@ -zDX =5Rmz6wvNћxߔo_2pڣ~^*(Y۵0V |͈ ɂHuՂUΙ2MYsM_+举:kd &O+!J}K)"yB `K,"-&MXڒj*Ԉj}hG.BB-~,,98VtJ#ZScDi3o&=!|ʐjpkKX9۽kca6qfh/͸3 М &LgZ{_%Ha/AɝL&% ve+əo[b7eoآsXbX*F BaDS$McK?PDjN9.bQlelSAD;bȭGvf>'4v)?H)/o:Q*C@e RR~< B`$:PLej;VC&͙  2?.9Mk_ :Ax3hX *ofã2ŽӺBI~xnims᷃iO_,ϨJs{Mm&tHXfvskcw~8ZcK*?/"+JDݞo΍mɹF2.g/?|G0"$h'8gI$`T1Aa/B Z0(HʘCcNE}7[<{7Z\N.%a7J{AH"dDGQx"_>Z2 3f"Q$(b%$ %Hq'JPX0T(AP$}-x5_jZ)Ǟ 3+BD/Df$5z|]gޛ{0?Ks9"OD(u] g`ClO01>l^[@߼,b1p^Ϳ䱧~ }:V@5"G~TxaË+EOYL֨A,,V)}F%I)uZ^6^nls``3@XD$Q,@!S>99ȼ򾲙P'){1eޝ+n呧iX Wq"l^ UR4"˗($Kxyi٠D\8/2ض%L9V L5mGv'0+SsW2#4ſB@<'NUGt'1뉼<֭n T*wj׽J!]k AYOBvycʖCQzcY^(U~ ]qީX ?^Y=^|y rZmW?/fz{-)PjAT=eh{W|٦^$#՜wqEQ{Ǐ$%ϕ^mu|y%~#R9GOjpԟVÏP~=ןgEzKڋT/ I&wp{Vuo aGa[o2ɽf2MZO_'iʏ\Jϓ |k&9 0R&+L^h>./.e}3sѵ JkA;qNmZiv.ͅ_B)Ġ=̬Cqcgs6]At.KPƋ)tz(4HOFZSZMqk5ŭyMzL(VY9"b"Q 2"f\'qLȂ8x4A1cHSe2qs9gixW}_t9~0Z>b&}Lꯖ;;Z&KOjFhRjRf "'¦{35`6Kmup;bFa6✛ " vqI`~1W_9sջK4G\DOgd_0u(|a' |8m߳OYdBp/k-sjҥQhXD? 1G'ޱ6\3ܚ.(9C6>$t`A'O vhjc!T6 yë^utQc'4N]=+c..⡌X2vbZQ}GIE;x?" z/AXZ'&ElF%TFHɺ+VbT,Ax˪SS5O"`!,F*WnR=.'85{ X+N"9 "Z9[<4 G͛F#}[+$VgkGNq£tE@Z08BBƈ_4=O~'+ٓp=t =vhoUB7k,'2O81BтÃ4q 8Qb e5WQ8 6XRDKMr_j)b3}7b&*qK~}|j {A;iBQrP'yK@ @1P1#&DZjjdЄ 0&fET}tt2FL\ |g?ovV0`c ɩc,ғhayt72Qli%d˪S%GP1 PFE& Mw{Gg9hiygw`.%?M:AnjҶ: EO`h-t K3p,uLBT"  #Lq46;Il1Q hv+P6ai0x*AWJ"FsuvT[ @Ad.$YZu ,ɹv%-ka#v <o/ Ǔ`:XF`s<-H,˕Ȅ*85y|/pH$#ՑG|FI.nv,M7B\˄V L'˜-6T xi6Wu}Gc-LL٤xri 7 Xc`c|I7>t_?? D7~ؽJ2* |lLl ưBň ͊FQBㄆA ?2S,pEJZ}ikg,+qM隌 O1y@BoTmp-`Tގp_3Z_(ל9F$<[\9Vpf}Vx#%bkilQ>LcI5r" u}iWWQϳB:mRѝ2W2ƹY\ NW9);ٝރT] d Q0&".'쯰l׈6yVp.ׁL1\VgD͑Yc"bj.kb` y>]Ȯ@S:[wtTAgwnIO-$Ltlc]E)7{uas ѶRX;(0 -el 34aizIA:Relg<3wa;SIFbs81@6 sXiqbi0k9 > *u'Z*P*N(p&Z=!\2V{FJ!ԧ^cӆnH!#D-&/UpVCIf. _="Q|"mn~Fb?'Qoq1alY$&6BPGUݝT̰T# PP GxX:DEk`i>R?ۛ*K=>fiRdm|XS%2mYk&Q4jwE9mOo~t/7yոXL>=Bi ߶h8Gm3#HV%-kTBq\M%q#63%z>A9;ǖtĝ&*Oh P&*GMga0EN|H{gev|b6䔣ȦQ>@ٗʶsI j'륙ϦVȵ?d" 2H0`}Q!iba*ax:y,s./K#תqL#?']l‡\Dx;Gz^I\4Gg`YgRμ $]Xc=|' ]RtuЦKK{ï*6KEjXAUfGXN%-'NgxT%l x:IGIaRS Bdd{e R@z9^(BV^zSXJr.ni^z{ۇLiqPY~j"5KEQA[V(G.X50Amsx XoՍaG8`SQkI9&D庫s b<0,둋a bnufBoGd%h] KdYu;%+$L#)EfI`*`0s&ts}\J,n&r$ˑ|w#>dm<1ؔ!ckL ĕQFdd<445A,aLQP(u"O;DŽ+3jIN`jsW|bƶ$dPNhfϋHn"SFٻF$W,fE}}i,6΋B^բ^p (%){Uj1-9bU1^7,RBq\rfs~IhΕF,j1mQX*Ek"LHbފJ<3#7_4 -d}L. TPVIR.1\Lg-_EB+2ƈ,1yG\[jّw^؈vZm0V0K,Gǟ4fHYT7!XrCv·8+~=OY,B3ݠ kۚ9,1z{( %EP8zKIk|8RX(SEe6)iǎ 7aBp^@h:+6`JFDoUdZX0;K0X)sŐ6H(G "f,B =3-ݮ~0#ܣg0̛D6WrjpJe13.\=Asyq=n `)7OV0fyx\ή0#k% fyx=c3,l}E᛺ hոN[-4IR'xw/U8_QdnIȞ)a,X(a- 1A|feȳ#Aqu 'oCX&Ća$paT-Qt/N 'll$q_k+%tIHݥ9`P cD zz4_kj& @|MA 39ž9!R.0rSyn"UA"T尚ZdKx2wԆkę9{lvI}{\,TJ/ `9lĶ>JdXw7.`AZΨdl^aajm>g^ YZ")XK%ٱ팺c߹SpR"tD1¥99-ՂL䨑 JH$TY~zV4[͘$V. 0--[ȴp瀿#B&Z 06=Ug 2NsÅu/> I#'dc@,6n5aaTCHa,=t\t:nUr;[c5U}VsCBv]+ ;|R^5kS)*nsQyP^-F_RLtU( u`-5jrR:0:zW`"bK؋MGawJl2}}_#}߮+6 ]+EaUMX=A,c锈@=8|p9蓿^fWڼ*bpU Www>cq>a9MywfYd_uhzrٙ*`S6/UkVVrUt2:ƭ ;"8iF3_^s|Ԫk,0i3&)ά&*˙!Y XzU|0KB=a,%eEG%r걶 ,B&OlbZ?{O?GrDEr!iW:'WXX]嚼hzLӶ\@i!y%ԩRc"Q-w+8B}#\ )b1: ٌB> OܞFB ch1CjI8E>Hm(Q"oy"p/SHg 3kwgT:co>_?ioR_t=XBb`,W @eYs A|cEYs%TElPG"[g)GhPr,ATKkkC֞ȷzs5yq4oV=d(H: 0dDc&R2mƝ/7:}ST"-QG TN>L*OX-'Z33_DKѓ}86yo`{U mÄnE赿9&VW4XJV-V])3!{45Vgftڣvޅ{-8AF]Syl|4V۟-FoŽ,_g< >l1v׷kwo<8<N~;v׭Kғ?&%*[?&ɔbga0jT N;hSγ<]kVnmHȟ\D˔D )zB5L&,"Hrue$ŨZRZn0Fz5%2l>򓓱~fp'1)i 3^ -$jmռP\'ml!*? i͚&,Gڸś7}x5# w:R.4~4asRP]rtqICMxy)y.8v5֌ boCHgC(jy޶Đ葡GEಬǃBe )ŗe uYb6 r/\w2_-p#넛e3>$&_Udmf$]"I +IF$)[_\|-0C@%K-%Yk,0i3&)ȷ&*˙!;ID.rͤ&NxJd""{#N3#R^(5,Pr 5,g<;|k~[䘭K9f1lnuQw`bl08\8(&Qΐ=WW3eWJ|ywqSOH%ŨZ1QEa-749R?0n?7saJúz*"Z^2$W($t0ts$ARkЭPHb:E!ύShR<E5h 99,8sSV2RY'@Ar8 |9sc<926CFx+/`Wۘ}]M4iyF#~. |snH+eߓ>|?E /G0,͐ԛ!=]>ŭ~|7sjXnhO#ept7 +V,5&f|(-@kI4f=0 z.p0 ;vy+3?O2*b ,݇{ ۙjGT>K\Im[\J -֧SRD2bRIҞH2f`4H)nQN jpcPkfS\\1ܙ\)Ճ|r g F\*]e!(VrQ jQE; Y)~^=(ʢl.,/2g UC$؃TmnqJdv>uPhR f2*.{L'@R8 caa<< 菢8;5)=Ђ1?]_}EGΎ]Z#>TWGb/**']UMnm=梧DۑcT {-%R]`+rGH}F}3NH^&Ѵ-V*Uq3DbN 79FaHgec@Qa#^qF`n_pJn* WwLw_lLt@"Ԥ6бXvzdt6b= G31ӳt<s}m'[uΥrWì ՇX #H|xe9*f4$1l7w&z; ݗ/劉۝@1 9 |0c@Ád~CَTB⇧UXyQFd--5vdГчayÌg~_I=-hy nuY9w\:-5ȎĴHr?5kJrs-$u!~G KllV6)?/Ŗo@w s/?uHG%Je2UE)Lr@b2HR I^ sR2%M Pp| rlؖN;W=HJ{ks?`BNt(˪>#)??5nU{_:kO5琲ە7{Tse'{5溺r1mƞ/۫cgVrߎ9_Eoʩ_fl4˕\nvxc՛zޯw'X};ƚr=7XV)pyjwxzwRz~LMC9٠-!SBZz^n&+$EKCF/6;vzV<2A(w=Ղ ˏl Y7#7)V-0T~c UY)P,RrJʜERgăj8_27uYjGm;AП<5ʁ6, .$EKN:;~ْ~c;YDtw|;&[;r"%SF/Z{ڍ3T퉁Us Pu^(ADCHȅhkRI{h0'5Pz_"=_O2tn?[!F/#F|B&eT ܺ#`οot0Ļ:ճ,gSEiXΊ_ՏϢyEV[S47/8A8KHJ3t~{*Q0-hR R )c *$EV,xftsN&Gm -d>̓i=xkifp(_4RV6]S}AS–-0U0J»i[nn<-3DMp"`eBh^jB2)Y*JV#5 $VF.F00/Cm6!]"A!eQ <R$Y%YF cb ;_1F.FYwX_DoC3b1$D}!R@btAh0(`r!xI%*2HVnFÝ&J]:ac [~1zSh&&VB l;a+SbPy q8J5k m;aWTTf0{$ĒRNkX{˺*'0#Uͤ-]%9yn1C-2]0:aҦ$tBZ,PE[_X&l)Qc^GO-@^[^d$|4 ED+m+.3!fW??78߀ߵ!|#:ݮ5N[-vXS-ִEIMjR3c8QoSz(xlE: Q ~'nX)!h7_ lDw ֣u@y2v[j.$EG0a8QC Ab`#:ch>iXۖڭ pm/Sց$HG|H/9yOp0GL \`)p=>6Cޅ[I#<|{mC~tt.&DF``ϳ)d<.j ?Ԛͽ֢NA mmAG@,,@hJJ(XG.9*s) MR2&句1wI.YH؞j΁yo8 tbU"H:ZV$rBw\Mn7=VM i7Ȫ~nƄwT q&XT@RYdD &ʜ)ϸ\$J  0yʔf")3JM4f1B@ -3p6qJ}X-^/wɪ=y1OY^Kd /H,A$c0`^3B[y4-lQNڎX+k9au#ax- D>_e [:b?}_*[~>Z:ngsQfq }B¢Uik4"0媸.n9;YObȀ!u% |V|ANG=`ރhZ=ݣXw ép3 y8CsҎ]1^竎MΦAn#[7;[k}*Bbat6O$Cf;Q;'=Ɛ!Ek?uaC.w\o~:Ƕ-$oÂچ-{t]=CB"FZbŵZhz.{'mGG/PŐj3 %>]繻us30dRqTR)Qx|~},6\.wUai,qAt5fza@ԓш?L{_j[ŋ2OKVzMT[SQ3le=6GF:9Ғ}fH^pDbاd(狯&g0%8˜jY!HT6aڱ=ffNo\,WYg+߭X)qB 1I24Na fpeωCg 9Ogcv~ :+r&1"m' 36 E9ڞ{2τ T%S2k[$ ۴DԼ{PN 8,P;!*v8pAV'psCbcN\4.pQG7% 0D7A/+Ǝ`ЅAˊ!YH/#4膸4x@[֬h`>X>=j@=_ч2 tG_& "!~ $D Rj?֐PD_}1# S0A 7 d!=wddžq$3.uÆףq+da\[Z0ʍl0HAzl5 w ՁlׁΗl )@QEJ,1RNQf%Hs8afB} | DU+0*:_la/FaT.u]u!zy63Z"5 RʌpJFX"9!մNL?-Pe^ K$94-@&r"OiԂ "q:2@@! ZipcvV V>d_ouE ЫWQ޷TUH|_~K* nn?x`Yݢ~P?u <6l ?O?Rb\mxAKӇWp\ߩߙ%_6*; L?>=<5h<ۣщSFF'/x襒*=#HmƖ0 yCvgUvMwJXnͼ2`/u=ו"iTcG}̚;%H 3uQ۲$nЁ +[To5z^iŠ5LB_`6d@_x^/dت_"Q#-{u%Y\4\z[Ckk0Q6OTM=j%;ЮT "b=#vޓ40,Ѓ }[ʮ>gDV2}4|H$ń<~ A|!B ݸilYYiV`pS= )pM GHun+9wP]]X(2^[v.˵G8ms&63$Z[Q cGamuqb{}n>v;Yު:Э<{CC4($ n/}٨*~6&jmk:_X`lւِՏ?8GNtc-`֫b ` rmDܬpOQcfZ (q3BDt`'iv\ܟG̟G\&!' mF Me<ؓ2(聅1=\C8ttTbyi "z(4npؑ! AB BTbyN7& }]0 ,/w} 5Ā9ҎYr7U2X7ו,_-]/T }ٻ&m#WTQq*I5Er4vR 53;(޲wFxFh4:?g]GB(+{Kx9^ӘiqԔ>M_fSjypJ@jˢ6mu$~x3_i'nG[_~6O/3 2-?lVORw}OnQ ܂Ų768%kIvkh/]1a&]lK"٢(kD1{g;P(կ hNU &oY7*W0Ѻb:c9JIg-HVukBC~*zNiISĿCL;7[)`_n,0z3`f?KV#:G1XJwzem/ c%],wڞGm2tsQs\0+eE;dN0$ާi g;y'vOFǕzݔy%R+BLȚt&eD$t)ܓ%3tfOY:Yk!t{yf L?w&ͧ7 ϟf/mVTi$%$PhIn1ю;a|.{_=b莘&BW )Љ8QBNݑy>uCKEi~zNFfȈ(N+NLNXDW8%ng;\ ?x ]i&$PY$p.7.תarbRӢ'3)3FX$OW 8ǪUvh֤6|m5nA݁m @V-zg@6,NywgؿW4mͰæa^B@5$dCÆ0aU SӦa=D@vڋ|J3>,߄b%k!ݍ\2K$X+~7"֊l=H;nI[Eu찖=Ep֏cR3*w[(@ _I̥pi|_wcpTԯwxZq]^]n ̸z靇 *;;H%$‰-#k q{X jlVf~ Qf(WeB_1y唦ԦX0q3drS.Rq;]q(Lv+>׊X]me]O|'.u2ka @Nm1B)36spLkbP0*JR)S:[XV_o}ɌGv9O=Meog>g{j=/e %+/_qs^?!7^?(qO~L Ts|џ8ݞ/+fc'O;t|(yZ# :pGa4'!Fti??]ϤR h^&tjf8?.2>*}Y>%f1_>Q*d1$r' ^?o~z f>}J dMs ҥpQ@P8BgZ:],¹|-Xat7ՙŎb_atAFl̿[U^޻*S ' BwT[BlVBB,  đL e};+-=+ ǥXj]=X%{&6uE$s/^9AFr~h*_/s7zzY>,(o&r_AI]:[vFK„u}ˮN${߮#@U{Sԫ-m!j7!9c`- ZAlI'9Gڨ ;Վ;Gx-R̸ jNQ%hd5>hSgX-LJsιaX k,3tj|mL۹8 ᤁ ҉Ssa!bwFk6F ."CӺ훟qϻ[}3{y}~CصϵUR2YKk,6aM{14W^pPq_%Yl'ixWVW}ˋZk!0؛? êz`LʜS&F:aҠD1r q2s3C37S|;-#-"B!bo-bg* &*FN!KSdʘirn"e"~iK ֤A,|`>0;Lk >iKô6jwb9gd/:N^gol4qX[#tIU6 zFؗ ~&u~G-/2j3͘$vaZfL#IP-s"_(_WŽfjull=yo ˽(ٛeC?N͝tt7[:Lwks]!`̅&DTVbk4zk AH[7@1VXki7lЈ[f|ye&Z,ug+MDM o[׹r\I1 EAҁ1wd(Y*Z̤Bb *0E(Y%bW-K{)-4pAp r*y2!b+yS^?B |$B+_Wp +(L`K7wSr38FD¨%FJ SqGSyWhVC^wJ*o B 5Pq˩7'qn=hF3-.r2oK?~t܁,>F3IՇbZ/('/6āb Y+Am 66S"aVókF)YbAp.\eZrlVPKq6]5Sg$[K #aZ]'`]QC2e1y1bZ4%պ.Sɢ#"HEE$sxAS~}tȹRuz: eUji\%5MP-h$A!z~ : *0Le o14QJnjlb:PHEtJ&gB}-$X ,}jj&AgTL;@m˘HD|fr8A5 ;TU`o7a0*U/_/0Hya48eD<@ONR2MʝሦdMj8=ѲqRb{®h/J_ OC"dLQZ`ՃA`ׯƨX 5{ 9kolpo3ߺK)k\Ky*ѽX3(9.U#m%Ьjg@3NG k36^Pa|wLtV_nzՎhتjG4 }к#'νm5윿- B:6lfwVټ k`Al|)!~aL^nL4V$8 B#Dn5?[rTl&ruo4˨ƤRYVcÃyATuzhpnjٺs"wHnjmF+-16QK9qSK&. /DkRflµS 8mBBjFuoiV$" #4iUp8iȮ%UhDCKk٦3m\k8S/\ā>Hky%!(9nX[dpq$?3ǥJMāMFcrѽQ UNC@ğWˁ̵#gZH` P74uXH@H~OO:y6.My82tQQ]t0x0HQaw3|jA~i~fS@jZ5k,JZc, ۾;l}.My& ʩL4TΈ)GG%C:b:-NKٕ|=FEZ;d1ueK%l9f]d݁rAne!ɢ[ҎUy IgD,+7+/sw bD%4YAU|{$OYaPtM.*xttSc%{ysxaT Ïz'BƭZB(bl6 pU " 51'텓%"XDŽR&cAzd>2vX$B(>_+iZ?뒨r]储`sn r8 ́ h٨KbI>vR$(ꎥr B$绖-y!yǥQWRtt!Sy$k2qhL ,US#Y/ZEXW܉:FMԇ}lkC-XvCx+{01& n Nq;hA oէvx™yw{ZbҲM;4lf֧/wҀP:a %o>a8LVLҰwu]]lA׏F ]~}*]QsJq xH0}.]^qZ=mN"fQ,Ej-7shehD4rd FjƝ\3jR.?rb FpSuh n&w_rGG2L n 4NIY$ɜ1vlCcDt"ba_oGv9OQ#s桂yO#Ru_R$oߐO+ #>]O<<_ ""A4z!g>` _>]Sw? _pl#"5oMN(Cѯ~ oTTNuOt6>pt\KF%z198{ϲ8-h,Np#xʹxRo>Q=e}2i*n=Rt욢q+f ]>ǔ£ RI#p=%_/F\mzmWU^0ĺʪK!}l9n-,-qmMݛe>/y{`%Wm)Ank޵dٿ"`gr@# lvK7 >zdK+Qɸ߷e]dHa8\oS[(]IܷӱCAuv}{ 1Us&GRt~wFG!=S&VӅV6ZW#߅ĞH}* TǫU)ow4\"mw`F ZD[N5(0jf[ `efp}i`>Snq7ͮ $ ;"3uJ7ݺuBх&!`/3԰YI{z׼#:TDImV([Na1N1AaF gmuz{S748.pj`!^i zDef#cS 7kxWp6m ʽ <(ޕWp{{:ȦS'2>!hz`ljs6Nq+cj0WR{T`ZM0-t@Ji[)c­!>o*onz¸IVAzHYsxRk(b)v)11"#۳ղQ( j&bq}l=z,OZ Yx&z̎6_1j7kg]Ofj7e֛ٯ9mEo$M̺_ue`![i/)ӿ>+n p,W&Fw>Dk)TX]fi'/#GS^HL@/:muy]d ##2²?}NdDJr,ᖌ?_]@T]%HPKGx}5FӪ|^W&5>t+]=z(͖}rwp=>Ǭ (ȭ.Riş-`J0BLJwm%EriNUּ0ޙy%LbW8~_k~ !nK`_ڋˋq",[䗌%=r"2M|靈(:,VpoO˵J!}f4FJOX_b>3b6o!+f(, bi@A $$LR+X9=m˜K>dz9UT"ڣ׊FgI Dݾ!asa$9Rsy$~>+Zlpr6D:&ta0kʑgBj(m+uK*(çk[?ԉK@'gHн 79RBJ{\;4iJ-BZ %Ld0qٱp`:;?XxU|RuuYUė4*W]t.D_z= KZ1؍l@Mr} NUJ͠?袕 8DK~k^l&Mﰃ-#놣 erFIt~tY*kgl@$21XT;996@NΛBA$&"_12*$%x4%* e# t 񊡓L"PCȀ?P  $SBD!'U j~~ yB7Bo&(7B>+Zb  h+?"<i?Rei4Wj8F3]0y(ZG\@-'Ki$&缒.pB^͚"Y] Eb(F!D4SNx$8JP1NXI(TH : (㌠1)|knmкK*t8i?#CWU 1i)R!I?KB#R#, bPEN(ip)ߤQ?f .dSfGuQ*U-s,n~kjDB L6>E2wZ0y3Y0@l F_~~wbYN7 #Aϼ'W~4\f=5P?u{#!4 ZGp_ f`L2)8CzUf'xC!bw&4_a4F Z9F[EP1'c&D P=x e9jjva.gs-$͂dfqZ.G<&?jFU6_ezF^gjĀ ^[M&4W[l}9]-nO`,yvsǿ[}w9r[\_<g!;E͋eBP %RȢ0WiR*AŁr>y-OjfE^EԔ=En QP |0hU)` 261qf zy >5jh9ifFz1N!k0Mw../:جjGᢩwu{b)!MMc/U6 "]̾ =Cezp\:gzX¡\jUUԥe<%1AˀZ"gI ^xGOae!u|q2Hr*p9Ul@ S} h(D20BH%1g?_3athҥcZfv̭/h6)K[fog}&91w3)dJC@+PcK@)@@*XpLSJT< UP,%́Q(ŗ*H#+!_Si7㈴ ES!BGw +_CĚVfA79tBkkqCdoJB=+Lz*cڬGR5mM\l3 ?Կ?W:50]/Xm~3jT3M~f2ޚxk&my2?!^LfU 0o~<<]MO$%tKFfJ~VӶS-1!|sĮpí[ܠ(g?Q;Q OĮ+sׇ-b{үnHLT V5``K0XY1 V{jUnY X^3 6Āk+,+hh勅gcq]i56޽}Wx'|OJԇ1Y{e}ˢZ3ܮKƶe!FS=ʷŲRQ 1aQb{x@,+>S޹tؿԾ+1PZ {?C yTĸGӇ~%0fAPÉ|Ec" EA4P8TF8\Nk NAfyƔB:KLj58|)Mg CqH%,c`]u.{c tp z%z 9G E> z1!jV/rgE4Mm^/}\P߀wP@KYljs&.*}HFGU^-r}h-.70`FD/;P%]+V3M>N9Eh>,ƩAAAN;XVRzy!jBzlpNsi wpT t>;;;Rt19tZݯݮ2Dsi8Xnb48XQ\%lT8=!u3ށ*0am9&NsxMpzL&bL ~O5oZ Xr ҁ ݛAŔHڱv{܍-=K xCw^0\Z;s nhj /jܮ{WhhNu }>̃{J;rHFWBE'0npjdyeKmPk-$q CV|LA=~o/7J/AovbiTL\h; Kx/sv51tpDBPS}2螌IS30AKZd4Ubْ')ל8y0d/}Oe^0%xy%u%+s:]gqsW:ZL`᥄\uն~/d)8rb̩d嚯sʟ܆1R-h`_ˋMgQ8 ХlZ&puIdm]ЗO˵ک κ4`$&Ĺ2;/Bc@(:z0OƢ0K⇬3x.a (иEJDi((ӻ]p\Kqe6-Qu^f1Hy ֺ'`[!Kɪ{Xcao4 ۞a OVh[VVNpV-NVF"wNV;W-I\/;w6mA:.;/{Gb+ގK90cdf06L^'`߭٠7fc3CR7XnkpBΖm~̴¾& :l3,I_KefYLqD;juFmom8 ey'}oΌH0#li$k$5njpaQ7U ̊Ό^1MϗTpV vZPJ2\4Mm&M+4XIJ^BDFdl)13r |&Lzl8xaqdOٹ&5`'D00C" L8D JP!L0d8QPL/F_,Q!Y(ϑ s)x.o !8Dqb(2 ŐG QHAS%XIzP %#'s,yJфA?3ٟgfjϭm7| ׷WH65D1`,:zl;?`,B%;Xӱᚎg"PB7F/uTFK=+mFFq_GKVWU9E%埾ztq/6U:⭷,<$TF b$$rDen$qfEFX'SeXHm} cuԳ-U"oYmҶ\9-p=XNԋMc41ޓ:R)@P+cjA* StBفJ /Mfu`p; ~tg!>*e~'r: :-L6rK:>MN/tr=-mvh<b{'eo97 Laovk@s|7dK٣dBd>LHsmdw0FĢ.?8u{o%ڔ}fk.`;%8G21㹷ğoT; Eο:%*xxKx$G+(XINnb1X# ʵ풚JT@6TSuC>)iq5ԿO\C"XIXG'-~qV8G]rbJ܃--9 !uۓ:qM2<,:yl/e `! Ii1ilWHsI3k>XCjnWpJkGΓdy c85:'bowZM sN%hj/+:طYy|} Ri6( ,G ZX@Қ kGH%UWH-3_7CF xaFx$D# ca QEGLV!{[aj,3UAᅑ0j-5O4bw\mR6c|˸񅑄KSma>FA :ɬFTHa4 VFh4{sj}0_ ֿfDeB]P7{?f UΥ +\Q+!v=Ju]ɽr@\NBQH'Eeڎ% Jeߕ9㨬&[0[ >[b]ʒI .V~TL=RM(eW,Yce-W %#{ӛ>.'#LRF|k}+8>j*R .tEtMږC=wu &k/IvU?d#!fS4@ LU&(3.i/veBڄq2'0^5|I9;Pb:&#r$fzLW%.̗gMA֩Z9|i8~bQwkKk.*ku6Q @% .dEQ$"!(TDۂ"%,iԃPq% %!WdG Q A8F,QP"&LA4N@Lav9# !l,H2Fh엾AeЁ^\/z\#!|,f'q""$$RJ(D1NIBʢ8 q%"G !h={{`eTeq3Dng|vaZo96)-w(?feoy-`l )d}|zr3$Hpg?w~)ÈxX#"ƔZB3IXRA9L,DJ,Y8Yd(r=/H#9mjb $*V`fϩxb}$Ƣ$KR!P{mN:z,V=g83-Owt7[K1$ДyNB._ҙ.Ol[H"z4o㸽S&N: ǻpnpeIHJ|LfGJىMwwU^mtiG>yuqs>SR:֊~EW>xU3~38݅ "WZǗes\_&ӴIyCj5}z}Qd;otG hECڢiE3[ёm7\kh%|#"vP|-i̗Hy9 %L9X) o4+;fOb*t0uot? /%-.cmYrϒ`ܫџdKB_{fNîxOg3^y^})[$wΫHaG ކ_@>杲ęןUzcL0޻+`~`-Lz_KRu5F]6:p'o|IB]/DaX1@jv޽XIΥ2 46\c$IE$ I 9 rcFL8A R` $R:+/fFhe3Z1yan1HsA5c~9ABCB 3LH@EcD (8H1H$!A `1@H_ch$_hꋇ8WȫaMaJd'XY 1W4}bāH2"ÄŐ5 ZA}d&qC<#M2l4Q굶UaTD[ f9}o-@&vb zBܙO'ї87$6؆Mȱ/npӽQ2ǻ7M4KjQZX8в4SkcBUfTz,.}ۓ&r; ptnĚOe$»K YvvC);Hbj3ߺϊA`kmk*=A:?z%? ǐ;VSP;huPr>Ɖ, F ݴv}ڲ-tҖZnB-+Qݷ x۴V"U]/]+9'{o"vgXstPq̨B^ʮkPTVIӰ oj,[E#DZZGr&*NbGhpyCh:'TZh5#j-'<ۖ5w'N\F3d|B"rY>FܑHHPg -!HDd~Βha[Y) =xE;?dA 桎a0 -؈L%m,ʥ6> ! f0YBBV?VvmTC$?P[ẇq"ȢpD~cD#AOiQ(%OuiZtTk-Y[eEڢ/L.o{].X*M: kK,nm٫ӎR6hF_3bϦHpúg7b*_,†Y|jNd6[Kj|)K1Tj=.U4?*x\.q'UѦ)m#GEݜd ;8` ;D{ْԒ"li;&2[M>XU$Z\wO7[.'Э<ˤ2k8e#@GA+FKt?Ҍ-&&R!7PPwtFC~60sEZ_q@:lƬ_8 V 'x2U¹<[:U(@8_J!,D>r:9˕'~(ݭq=3YU3У*,!؂C8T(2 c}>[BaR4O՟O<]u}#uʱ?%MDa?B9"KNs$F8T4f${\EaebH8(0Ic(d& mjeRPݼq2## )Ō(Q`yAP  pL, !e:*FTsUXqZ* l` 넏%&Y>/gY~\ ;:;zRO9 -O)8ͰМJED"HKI$RBPG"jCR c&1=J1N;۽tu5gO،ݴATQh6.񴘫[ \[ڮ]7HR-G4)ٶzVJdhQ?u$j7O_F|im[H/T#}qă]fʒ2A=c-x |A S6CX YAspSݯݟBZu=Dt<ܜMyr__DYLhaHRde06ұ~..HIFx#H6+E*MR}-4znQ~(# #NVE$ކEisV,I֓`Ϩ\fD DR^h2~[KZd֒'+PV~寋^:C^MO;oyoq'#:OJ7Ͳ޲,qD#"cCЦJ zĚ){=&Γ DEJ0ϠDڳҾ D<'yP~() I ,+;8P,֙p$6L~՛8$ۚhǼ5cF@@ %HFh#Q?߇Մѽ#N'6 >)8o/Zv-Gl9 !~`dr{p D#$жWDIz[x-AlyEm w펓5?}4?h: ,3݋]XK_s50>|zK85Ζ e2tۅ&_Տpo<~SWҍBNur'rڶRrveq3:q"Oet]LANOB;^o wt!ZjF-1nv} +YqCQ;=J6/:8PߏcnM*Z[ StqoM ~*k4{kd|uY!<L<:O>fZ~K!3(̒Pq0%Bq5mvNHN/b؃D218g} F [`0n9!҃J@Xr6[6JӼ biT8wp֭SZߍcnP(GCӾCQzsN~ 1zO&g7-[D"z x wDP8 S;Aq4 NQv`BB#Pm8B3$Q6! L1o]~ǧ`f kl7.Wy}B9 k '12 ?t)r(td@AYRc5Ϥu>H.X ZǪVޮ絷k6VꆯX=JLDTaih/؃0$F5@"!ѹ Hz˩0c "^1fa6@Ɓx+燿~-Te@^] 3Z~+e0|+nͺ<\?b~e5ɐD~dLTb;8]zn9{$rZf_2eXWx:]-×>UhTYOǥ?zO0Y8ntb Rvaz?`3ӗ\+buߧU,Ţa1q & G'cH#`e5pa&ՈfUG赔Ah*ߡ/uT+8 nYc5-Fps{sj ?띜*R7]f:%j2L8$48$!7vv-H+\k .Pfe10Iޱ)ji%ZmP_mڣbY\#M9G W CE\Yw*?5(IwxOH; #mڟgvHo[Y*d~+deo D?WMyl ںm'q婤&wFuy.BaI&;;Z@$I]yaIJw6Kr/F䖎72fj;n^X¸u^%AKʖ@7!)S.+?ļPڀkXj/_E/ynp왕LzfBA"LUndwb=`AX귞<W>t)Q wgGf-Fto I.ժp$Va*؏s}r~3J12[{;!|IH$("f؀1Ox:^|Wl'Vhl+s.dܮ+H(@TPRPT$'sDr,BYM?Y =S\:3 M oOf_K)E&Youo'+/ulلRӼ ~D U@dhR}mW+|64>ɭo:8Fόj_N&g~C"{Ed%e84táމr!(4X8Cu5(8S89ܠpO! qjlkH!VNqO 6W!#\{}n%HFi#O/Ҹnru1n0~X_]\ u(!ݿg.ٵ5ᗇ9KB,~~>#-R~?VEy6s?TϔחlL놃d%쪊'WT]KIcNEڿ1ԜN)YMqR1?~jbXn>1KFs5k1!n/ ٧xClU.:%j6NpD\KY2C #Si-ArCJPQ A09)E^R\<|6vW $2-#?qVyHR%l{ikX4" #'7*S=ފ9I!mr7X-q$Ra1 bI"ò ރqG>e[ιr@6/Fqq&t ;feH@{#{;?gw\1e}v{kZ(Ӯ;r9˱}fbٞe-޻oǨU%1ɖaջ\H,[7/%~a 0XOc UvHBѽ ^ ӆ0䙧~gᥨz`Iby^X`\X 2ӇU}0 R)!h>DA5q[X@LY'ĉvm;TXɮ'LWHĘh;Mrf3_k=y(\b22Q&eR@r-əODp%9kg5U2=TArTB,I() OsI*U"ɸJbIef5FBX[qZ* lR`&4(I KDM+439ϣAxy?;^B]:%x$j{:8 Nd"3!d^,Z@DD&P3!!TF5M'WTϟNH@hk&uuAs\O<+&@zJEѯ\ics𰚰G zcX=T-7w7pX6ϔLb&l}~Nu{fH2ߟ<)#(A7j~hT N -g_0c8g1:5 vEkwd' .z_jPѽ+8tц6fx]@X۽`Iks nP!'65m)B/|ǫu6n}űrŒ`;fli O @sw.qhɑs Ul68QUi4)$IKN!9ZT8$-9GVD&9Gp F5}Meg8v< g(AMB TdcЇO+YK7ٛNVn] D~-XT>QD&'v cl F3n^!.쇗g<71Oxm]lMt pTl>px>gWߛtlrs}#(3ozfᤷyEu^~Hx>_Sqm%^O Vz6϶+rk?}{`G!;DMi Hb, 4sC`8kov>y!{33WMP+% C)τۻµ`qo0]/׹ p29İO?f+>׊PrWRjHTŪEgwo}O3x-[q0¡u$( R\ Ft% &ju /y>(g;Xӑ\s3+v##7CYHND{oBdnkd㿿~k/NϽq`gFM7[Ed6vMOK]"4q8r#qrz30nTunN8w$GHCY6muh3P B<l,|_pm (X%Uȫ(ȪpQrۋw"{_zHzQqr`*j 14 &Iۼ>SY|{'\{I,va FF t67^^rݤela}, 16d\sz f{nzA2W&9YZi.-vuҍ Ԅ Pt.}PF9X+sQFo  ֱ`trdaj}Ū3Ń`daQH7]ٛqdnQ_E/J_>}Bs\^^W_0H";Vh!§ ac(S!/ԩ9'e>y8,k0g0Y\݋X )F~||-QpX~O_.Nx)'ֈ!OCOo$قw?{O_?-;^{ G# C6ˎ\*rRo`}TFEW蓷qC᩠ zT{=1|^w,KcYJLivDzz, Xxh[A^e߱'+ B\XWW7ÓUu4XU0~Ȯh|͐P??;\sS 65#7MgV#p}6 gFsX)܂mukv%Z: 2\yd0pg'wϣD5e@|f&&̈t |d~hۯ 珃Q۞_}˞ *\3s|oxqa̤Ao{uαۜ#Д8c A'Qy pPiC^`X~m]"2Y!'"4O;ᗄ;c_ F VPn]?ܽ/noKq SYҹ(}!G+hH-LgDMR̢K]{y}udb3Z ޹vz>iqT3i:O n:#A yT^rw_88D gKcK{6ĭ*:pMd=&L+(XMN!&HZrv 2<iTF K0T$B:I85R'Vsjr8K$:5v# [IP4_q EHF8j1ul+*[¶_[ũ,Ӎ!.`Tv)Io07}5]o՝nmۭܺ:NHQs㚳]ges簠cԗ i!T[0K}Z +>y]ѪXôƪ\ԛ=O~כeKz-242:id }C-ږol5p]UmYdM8`C*`n׍Eତy]Β*\t:x=iA4´EË J`q_tTiZ(n)rrd6yN~||A9ǚgJhQBaz={,'WDejL\. cD$3$!,A\&*R^L1 OC$qeJ~sE^tasJn䎙2hEu\ c#Lů j!9Gč֚| OhYŏ{oL U7ߎ խpvw߶dd} ts6HW)Req}U804҄C%C >L!";/P"濘P1}c3ڷw?;KQPX 3,&+IJt.g{BFM٫hKM*>⮜Kp?<7'}`cN]ycΕyQyW^QZގ20㰷c #͍Ђ?4+'.2%Qu.pcL6E#o+6x5eR AĶY O0 *5Iq J(տc0xc8/bCBDRA<Ï*`?>9ɫԢ+qm%^6SY_ 7 7σok b(;9[%S@'p f.H_I5Ae.m@d. p UDR}UA p*M!$%U7@'d+# u0Q#h("HLl@B{e<"- fRI|YP*SG \M_){&Sd'qSd$",*L[z36kj$ɪo`CFӥ ,^BvMz@*_׿^`&ߜӴr:YK9W=2at;4q@xx% (K3[AYbS#=/3p ZIĩc$13C IŠ/zd:|X0(>Q @] `4~"I!Hi-9;gK4k? w#)AW+:Gش{ #}0*c-xe'A$϶ղO ʣJHYBlHةȊ FQjߓY3Τ J+]P5=FN`3Tq=Stc\1A,\+zX؏Wż21j*UQ!"YhیWBv+W/J[{gKA~!rftPд`݉%l)5]J J}-5C9=?6k dn$Koܘt>/6C(^wiLC(0T&lp[_T}Ǻč}x46+i"ùb{+](ێNHlqG٦ݹ$MϣrzjҐ?S\bAPFXA)3N76;/.}F8/Bl*lG}34IoLҿrqa/=qt=\׶׋wo+~l.FLgf_rK&N|1?Y@%NE9w/83$t4h3?R]Ώ(i[:PT~aD]kRlă&GhO$'=,SQg!ZVu+OO%z{d+i5ٚZTh WO6 l~uMqo^dVIXp]%p^JվyzV\w~@yTF[Ryl>K`XO!_ P00aF5G1ÀܑxrE"lQs;qtGKh9 ƾs:_أo9-Uhڢo6B3rF~.w+m[{-t}ogMnP/n$ !l^Jgbc>/҃5J(=X7J'`aJB3DP:Z1ɢ VœScBM 0 ݶCݠ@N{]̨bּ\b=fdPQ嗟:VȌ ]{!(PbUqYȅTLXlba슉0&*4!Xp0Rs(t|Tm魤% X)ܲGD{>"̍G)'CeE8sg,Ӛo#iّAjQ,OljX5 d::n3 Lۙ18M,m~P5ClQjIxccy~ia8[[IGqηnllW snw]qO[hZ^̌k-g#ԓCR*Nºk1~1狐rR'ANrGB fQ9UJBIS)b>Z!*dGBbkMVʽ˻ &ry.fm8Nͅ>$*;cRxmfejN@P l S2UR}]4Xbh7ͷ 1"yyEyf}3* 3IާLa,gc̋U#v7^l}zN囡S5~zQl]7%=E @r.EwWcxkvItUs֓Oչ'(]:FV`RW6ZxN. ԓC;D)X:g2 0 (UF29C(jxJ%\X&3b!1ijE*Ő:.GQ[jk-*kD|++h(,Nu(D4:yF8E[ XS){L9'EH¿@ ׀ $`𿅍M5CE&-j裂mpbjE+R%Uh'vE2UM$0?` =Ε҅juYCBOyAsDbhRafX*2BeN%X 24a HdnF7oB7-%h'J/LLIai(/IPꙀzn7kۄff=UksA2|&$weQ#eܥj0 *'b2pE ԰L{4 x=4M:U2Ao+xbQ,H%tLBHi]JSQ?B#k4[L(h<ʼFkj\G"jf hTB:1ť"}_U:)ZϮ fWD5j*q".8oNU5Iˀ y~Qp}]BK1oPOݯ]u[9rAH3*épʸp!Yfa-"̸akxQĄH=.h{ME-T=@]%j/5EH%b>[ UւDt0Bhb i58tr6dc pg>Dc,[#y A .%.[UW;Y42EQFg~?0GԲh.N(;gr0J(T(X=0 N/;W>j ~bvebF]:0ih u15)zםOmSY1PY[n9a| J(Bt Gag;jbILNgnT`XC$HF&ifNT U;MFE 3]~e(:R\T~GR=0__WC@HwCWM[5:AJqka'N kD PZb5>䟃nye0vgD@HQKpk gۻxwOwgCeݘvgcҊWaJTCm Jˣ/#SbCkK36o=(f.V$Ɠ ?\u ܫ^q+Yqb0C~Myzp/!>gk'~x"` ~ETz]cc:_,`6ɂg2> %/kXF:߈3$ *V3(zu>K_:.ٯ=2 (ŖpL9Mbm΃D9,:?!dH;*Dd5"Reogy* x> ӿZ#v Zg[4$ג8ln߃?f 9Sk`R-\eNL<_tx'JuKsY¤"N3dq5wUW+/FtoQ.4&[=ڼd#qB;{wI1!ki7V7JoYSL6 C];%7Z )`aB9S7F[):8C\eR9AJ*2S)5!p n~MQX*n!S:h5?pil~/S`cg:3@sx*x06?,/am34$|\I d1bIE0A ^>K2U)\[4ohAkH)kB 4ԌRB1 /왈15- ieOVˌ1 & Sgb zm8=vWL'S"&l1q[Pl:2ձܨ|j9 9*}GY;]i]Ɯe<:7ޫ J6D1C%SW>McURƌzKٶ ">v p>rxpu#<_0߾?8O~.}W~J?lﻣ<닌Oww|bߞ9?NskY4O$|98?ă2/Wϊ[ם9vo縵9(Go[6Jhŏ/O~?ϺGOKIxɛW˯oO~;񗣣}Mvm؟'bTYGBE>ჿ$|a8#4 ޣ_߀Jy24 KJY65VGyxk|q33\^?8}=ރrʄVէcE&Yz%eQe.4(},-_Ÿ &~'olk't௧ٯ^yYpdNsV(xl`=>HUnt5| ©i4(X@s.* x<Ф~;(y>+hXbZ Nƛ!/{&Z^< .w .ȤArSsw/fʩw/FT`uq?(%$O_Ϗ97(}NL`i>ĚAA 9{y<́~6gM~all8.?v  sHm7u=[]IP.@.mHcҸ-0L"|L3/ \9c@< 0A/lH澝@+EuGYnRrN8}3Bα>?>Z\h+01S=/x 'Ԃ_ܸl7\q1 Ղk@-p翾eBpILCቡi[]b<"`7S0DybRK)axTjwKcwwSԵ$|)we͍I0e *?ZƎǎef;p SI{I*I /DYUYyY+$Lh@tK&ұQ-]F~2G5z(V/|!I=kRg B$p(?|qP"ؐR3d8dgq j‹-qwȨYT<2\HJ(VDqLqo_.gDϳŊ~]x3[o<9ޯ‡|6p>_(̩(*RQCLMQUzTWկ;_Zu y[J:^c_F#j0.aثuE2{Ltbu6s7h} =?&jB}ɴ1+VȞ>`ꙮ1јo 0Nǘ\i֗?8fR:Euj.TE%뗷ɕ[=nR_SxgoNnnWR6+QY޸%;w݊?!ii?ʠCIq3K4݈1ә2CS02-0\D6$.c& giFx^Wz cc3<}[i||w?ß~rwq2VwI&C"KKll2u[MxٚמOxq1%3w.Aa"kt" &m͉V6~J;2Hxn 4u.p~q5MyAS˃k5F8[ŰoTT t1\lVXl 2{7g&[To!H;^;srї9n &*fK%u]8| u@?IƐ|׃:[b8"H"b yΝr&cF>;*=!.=m#Ų#t/5<Q;^\ ԀEdofYUgi""ĘT Ā3PU6")6\V<ʫCSd =og`2Md'aBNI_m%k8Bitgiv(}S)T6DpSȶڊVj$}:$ULO$Z cRhf uF*8w} mXu)dQ Fmj/ "4 so|0FrJId<_adCҡ/=`yo$tͻ%V=)BOQryO7w/&^yBhRtǣiB*+aF_kQdFx99/(VE2&$1CGds6yc% L;މyY P+Ω0S7T81SZE+ͩplA)ѽ预JX8 s3 13r]WbXMDXs_U$ЖPJLIլ-$7R" u뻆1Ĥo2Ҷaca mFذd!AioCS+޳rR}%(k_s 6[7P7 H\v;D:V\$Y)aqQBg2ʙܐnNG5JFnȓ]'a D͒8||FdH,Ml䇽dR8Ϡ<3B!T^,0h}f獈GivY獠aK#߯A jq {Rt!m%j+yp(>!w*~)© 6zC)1^+ㄖ-+riL)!3.HB8Y\i"VȘp)4Ie༨΍ ȚgaNp()zKJeNJce5ܜk[n-OU AHL1j Mm(*Z&E\eC-PN(B2UQ_cbOv֓ESL1$j7j齉diKwtlDs]IoC)΁u6 PinA6dWЭ ,p[N["|jѱ@2O=h5s-Έ3۱̔!WeEGes7!ڴq[DɀC8j(npi7fT&4PzNȸ)RXx.)q]EѼ5s7ze2!x Je^3hlrЩTiSq& m\f \9x&x*[xږK=. cȼL 9JIDz$ƘMӜXn%9gI+!kt=YI]71RcT$P6Iw:~NG6=D5I ;8ߜi|4ϗZeA]M/#h9ǟ>82ۖu;* 萸zqUJE I"vaQBcm ?*(],A*ӉoeTn}HK 0)wT.3(_4"oG5?o͏t14Of΂H˭ͤ2aFIV2sf $mɵ&P@~n`+2ض/z?PO$Hߑ~Yxm?wNVv,kwѶJtH#>GWW}Lɭ:"&Y1dCw@,_O Lhߞ|/{ie3#eVtדkhwA.Go6nE)鿏^&}ޚ\<_wV1sD8 [a*ͯDm{ eb: g:7cqIg, p"6J- c Oέ=5]xO[\QfFǐe6%ߴLUlj(R\ nf,tٶIA{mknFYp45[ ̀yeCbl|םH:r!f yq=IJzM{ptd6siWƆG]@yY|ƕXG6-?a@@7g!@=13>*趼 6_BRwt[ /Žۢ` J}nf>K{^ ngjInNb>/w`I`g= .5ݍG͋pQJVj^ۍ_0<&d3.LOt%fv,UL%rK `˴1L '{nT{neR* _F. _hYxnV e/TY_xeH&i.$>d-jk@y(J&7sATv ,tlrnL:IL1FH049^Wgr<4Ff^"Q+HI kaV * I@-~AE7,xa ;P|trl[$ZTTVj.zKvү!l~ũ=(~eS`nVus gom+mo7GHyT;z+=(*T 75!޵6v#be6H][ ebg ٙyۉvdd2[[^t8:.P v;ҖՕ%}7YQql=s_!l YI; p/(1że@]}nJje=z4exVfx3񸓫܅c&:4p[nszV9u/F*a]b2Oܝ,Ox!q Eĥqt^|1dN()?ȃM(^ȝoH|2jI(NuۯL0TF~dK[ ;:Q=֌qo挱kneK] )YQFۑ ;rO)I!B.'5tQ\KZ(>[%U:cn0\0ʯOZ׌E yht$"N_B xb5XtrԻ栲 J.Q{ɐ3ixM Pj}yGCfnOG%Qkb0ͫefEz7z 2_Y9oD%EY-{Mi6{ >}z: >ڪ_RVm費B0ÏH i{6h$G֟;+-uĽ*w6h^t wʎ/j\½f*iݤ\AֶD;:h%yܐb%T> ק.?dq s`F4LO^[rnHAJm~)]^.ti ۤãWX{ĉڵvu(mh6ͺ1ƨ<ЎQz]w~XiͼvùWo8 qn\Oͻͅ8kh(md2g*V%1fKxR6sgu"撀 G R!;tVгF` %sx-mP#%46LRlvMBdy7 [SXyi{iQ{ރm3Zpg.m--3xfB36/9%{&fY^=г۫.W;[R ̉>ʼg֏OLϗ4'yy8/>?| ŧ<{OiD6Zɭ~zR\=I+[(JyuV{/.)l E'r"1mqy@/! VdݿFd=ǖZM}1 vPHF #%KcH\HI!ī!| h%oWG!S0Zωo+ DA`?3`es #2c5ͮ)#d}(LAK 0P=&2hzC[a>2kR5ҵ5>ubG_]v1}Oj۷jaً̡;&Ne5UX6XL*IE%D+9.೉U퐲o"!u4,X(BQ#ؚ`])1X"f(k_l-~k4x}+x+}MBhkb):Hx,L= :3*E R1F$xMd^Eb23͞#5jBnF*Xd.M$&Jtfi!t.Ž҉F#[p.Kj}Ơ-`eNx'+{ҪGeѱwK%h; Xuzg+ Rr dBd0ڠ;4E1ACW:…gOiq?A { U}j0˹™?F` 48ѰnH?Cvĸ$#AB)P Bffe&JFjB2>e؏~>DfcjssYއ-d{gJ9rviL͇*J]r? Ny L{+_閕ߌ&gSU&Q\FE~Ӫו{%LPX& 6F!T" &]RԂjk f^z4gyG̾^yTS1{!@|?nM뉎IǨ*{]w1#Ez(HG5439=VLhMզiIxJJ(^10Y(˛"Qld\hT5y?S p|l|)ZE&atSyEBC{Oj8k77t76ؿ9 &~]BL ]#*X5qH)_JꔑȦlAdP}]V396JT@Ыtet]Rfʁ=2%[X5l2SOEUTu'wGUop]PFo|s=sx]šv7#=:bf1 qr~5BSU۵/Il@%H1߁VZMJȨ 1Vc'oט4ߊs@oO~(P'|Ԉ܍Ƹݘ'b4xW!t3wY1jcRgɿu6Z3ƽw>[}XNHsLltn[ 䴆níM=$ :^{,"62#}VX0[N/ZP[0!FHQ_C,saB-t*C-L H~@UWw+Gu i(uGr~bϷbٗ1a%2+ҽeeҍTЋxVvx'n ,Y[#B/УhtN걫$;D]"#p~Z5hgX"y< *BDBrA,kGUr̋xUm꓈2F"VB)dW3י A fyWUkO@:Fɮ)kcp+)ňEܜާdW>;wJ=;",a<6֑1.0F^WN@-k!P!pǼAs½q, ǀK<«`ô|ʗ~KwH>jgZF[R~#t?Y^C.'fyP U/_Q [7eI\ZM636"T- ?Ki]a `ŋI@(f"y%,jMe.GE&EF;{{'*RSE-޼-x8pl3Pt"Iuʙe 으o]+vY"Ps^泿mRObAX|F&f(ZKL[6gs$>alihezʘUD٦}ұ$$+ bZ1{jfޕO>>f`x퀢r`i+$_WpDńƮ0#~h3Eac2 cU\7Uxnh2.0{0܆u|w!Ff~K%+"B8UPjL͏+RA] p 3 e.]` ѻh,XV6kWV]XeGȢ#Lbïdf6oa1-~A|( yn,!c&/cdpp =, 6 ەd&XtGz'!u߶&Ѱx )%_+R!~x;K$L;VҩugHoEّswܞد,#5mK?Jw+ΐ\Rؚ M/ۙMhέv/Λ7.'ee[6Pk\I=X-hΑ IE6˳^T(g4DogD`L Od lj%PIq kJF̙/RA,R! E #ĥ8Ǟ#D68k8 Xx`+wyF tE9XQí^(80.bc?8PCs)=0TpS*Nnz%@k~Ӷ^Fsi!lC(z Z8`{.P둔+/d}K`zQB/Q~09r GV*c&Bɇι(҄Ǧ~?MTA 7K*J ]x(AwڐC96놜s: .Y%닮LkLND&-'1N]ol,#K"͂Gߑ(G~ڸެ>,gS{:"Cf/)DLkJ움a,LjXIFp9+%b0|8TR,s)lNyaIUH5ѱǏS 8(T52H B'EC  #4:<[5B%x7tD$!4qe'Nhzx_V*:Ԛ9tB-Y%<6<_nk ;Lؽy)F_n#韮J]`WG.c/˞f)J{mQDFOk^#ExjҼY coKBRbnd)R󧟮G/Il3SKrْٙhm-:ْgp6.CX̜cp8{5/d)xIО=cb"tb~Zc.+om|͍Ǿyzַbn|oi &iT"O'~|ܲ\"?~za lSNƭa6F%13gfffs~f>:d9@P"GoB%(Ha?m4.*8Ng*@KJs5AG42.aƘuȩ -2+B$Xל \H4LJ+ܣ+nTB6B \!D @Ùn|Q`p(tAü"h[D6#X"pؖm9tv!i4 Fv:֕_(0BNg;Ah3SU/ng}Vu&Plg v~HMQ% @{HT?X1ď ŸHXz <[R6ݥNR=B}qrm, eJb `) X u(xDBƆ5 ; 'v% Mĺ2:ҀgeNc0x 2^ML4#_t/M8U Jwd"$oV5]cEA{([QU vk ] mo:rU_BBڐ -1T7xʩ9 x</AM.~OcLU΋0ɱ^΢l/^Z: Ę[eAѰ;MWw>{w)Lc6k| f:琀PUh} ㆋ1/582['fX1K K/ox5v˲^,ӱ%Rݒhv` p yصh~Hޡ1682I,ꅍ΂lӣOLEe *c*hM{Fv0!{EJScտ\|:ѡf]\ֆQI?mLjW#/hKH,C#]g/_ z`DJG nhs\ZP4cɥ`{8 uM܎T`wwo@4xPjXV/8qUٞa1aWk?Ej k{P̩L īh)yӈq`f83XQ+Rh;>ǣ踔%_qy|!8w՜S{GR<UXTS%0^Vt+4gJ9Sd* 4' ֡T4aHyJEE& +$VCA֔XOu@21VH!򠒔IeUFJ,4=Ic:WϨ)D@AS#ā838FRRQC+ V ?đ,rR. DeΠؖDVPF= sJΉ VLbQ+TX7>Xր ԁ56 u`A?hY=DХ$Ӽ1F.i4 T N*9cP(%! g% #QSJOjkE<$ +'β` ˝&Hih0i ]trB ]!d@+C2#1r`:[(H nϡl5AH/ Pvo'"]^O%hZ?,iܝđv_#L/Ktl+(w_@"J z{˭B<`_?Շ#z@"` nMzn0xH&q35Xj^?ofOfPTing揙 Ff$YgWo^=}R\IMϚ7\aZѯF}X1؃~Z -'m$GWzL2WT;٧<]U=݋ Jv9uf*SWPb88S*$M1\5%KT!{:6R& []hCe0g i~mj_Ѱ ڪɴ֭Gp);DkVgDSiOZ8Hiq415U]F PN^l j^“#>a >J)#=xwz2 >%^ҟ)O z4!,^D=K)YN;FCxˇbCܗEn9]W̠gZiƝ&+mbqDUƕvd\.5h(zXz;D^ukGјwݢ% K =gKP Ҭo{h(4Uytr;L{榩3 >>Wxzb舷- +z5:A7̶gv,Ög% A1;%A{ cw% Ӊ6"KaoLJd)V "]go ^ ӓmG!-sF5R!űg& +y6J{c 94BV!Gj"vcFOh>vJ[7_U n~&mƴ)oA2FaD|A}"(D`c9B"A:J mkzeHXџ\Ђԗ:\"+A55h裑B65 TA [ő_$-LC(o_4s'W<9 ]ff~oX`> [qWae вǽu#sĜ}&|91mhQfP^V^{4)OmGXI%Ԫ}TP+%S@.DJDh洈 \`$gz |&gR,h$&4;ALN*ӞlW )d 8S[:E^c!G$j~?&) X ØQhϬa2I+KVq)ޛ}G7yLJ~zۄɸ 4l/W˗Is#=A[ I7~.RjWizO>Y{;.7ޭ4Jrj7)@_y㪲A>}?鉛L./ntp!>!߱ĺ~h91cxa4Ĝ~i5C`aZ[2-ހ5yb Bbڏ-3w f 9}іΓg@|,)Zt}'kW>91iLɭ/i2N-=qM- #Ch>__YhWwO^%wCBqwr;z׬p%y^ #£ 3oQGBIP#!=Pbdvt'__~%Pϖ@ULr9hV[yBmAmB{j"|T#@nF7@sPl+I+pgQtCjdTtBmAXpxK? $F. ^Un{C)$8 `ȋ,sZ5^xԵBֹVk=Ru#*=#fğ[45 gӆ\rg}@m(w`c"4[(iH;s;J{ӚuäʽnPl7mf#=8qdʭ9)#48?L䧔lj) ?1]}ߧlLJm Z({mu1gƑIjګ2(=r3:%r9,*3`!㥉i9m ԨW^.RkÇoF ю5 g\DnAsBfAC :VF4t&ɐ.t \Uf^d `=1Y =lѯTЌJZ?.w2&~>: R<>sb#WfD/?=>Np@5V8N_\3LAV+fs}|w/ǓW~"s;#[̷!{?s[ a ZQ;) J^XO$)(,aztsCK/g-I i0 fLnQDur`?a+ЖF_Adټ ,βJԈ!?{pЬ`;䞨YBxge'L₌MBB *.DX]x9&v۔x)3#S -|ZuY b4V}h8-d]g-3M5g|aKۼj !5ƕjkdWbfZc]؃]Y:*Nz.s[I\ dv0>ۻޕrJb{I&|xآo&?'o-:sNYn7CAtB"͔S˼~L鵪g; 3N5ʏYU.:N?t=0lnߟ>}c bc#)]v:=^Ѡ =] tC؜}^9 TO?mժk T` TځdP;, To9azt4++sK:@;^ܿ[al;ѿV{r0E7t::-C]:[g1M?_yEM̚O_X?>|;%}Oޗm>"1$j}fh^`R=c"N3fa銙o1/M.Ҵ@\D{ȔB Lۏ/j)|z1ԬL/Z8b՞%[blrf4k ꖴe[Rk/4rLJW(w)#@]%O0eOVđ0Ԧ̒6Xn䔁쌧0B )yh$ё$N`k|+H>rNAI e8\) t [k8lh/~9e*09Lr^D .l"9H!IMbCBi*ύL[zgN"74}vM^yess!NŹ4%xnmIHS,cP/ul'ҨYHwRrqRJ2Xڍ hL)YI )N _\6FZ~-ύalƇ4.zӐ5H{]cznҔQKVvhxr[ 5[Q_v~5}GC9Ʊ ٗ!MR&Dg[D]t4Ε~_0=72$ 6ْO12_D$[j-갖0*#$IS!S-O*܈%OeHϖ!5L! rڃ'ni(KzwH36jqK%o ֪Ѡy X j[ApB+ʌqEzsc;oӰuo^u[*b7;Fɤ͜uX3G9QQ4-u! UzGUo:+:noS)n(ĉ7:uk0 ި* Mt@g[f+-).Ю<%8$jetk.GiZҲ0dzpXTZIwR͌J!'Ig,$t`Ѯ$0jF!kp & h+79<:]!DUlR|QE1K.IB)Gsb&&IrVFDdVrm H(1[F5Z4 U% UR}ވ?:@3~Zm[&\&@Տџϻ#@q|iw\`eɜʱ+nlz-+H\krctڋp#h?6#+ƚ7`w9iefɘY l_Q^k% ɓ!ox ShΊj^hl$%}S{v2"}&dg/AC{ilJ,cw5TXfMVZ+=- %I[NLy x(Z76IF}X5 >G11W#qbݤ5RseEu1(b9#@਱Cb}k pK6?Էq+Ů`h{lmU^>OJD0tRJTYߧOO8"JTwd`P[e*%4`_t1]ϖ%`@o?#@ڵAƑ6^@#8p'OƾFC:txv2&+@0:j$XhN/"0xc(ohdvKU"Zv 2xDϢK4Q -=9ҮT';2p^wf1wCe0|оX2[z/ b<-W%Va1oU 嗣/>CYzMMXMv4G_RK*7M#Ϥ,17* LPڧeH)mu9e~X4/ZzK~}蜨բꁹt7~lGC$pu{` WuxͿ= rY>b{G%~ v֏Nںmܘ#~o*7HG c~$*Kj5x?ҝnHcP` Yt#9Ncg2[ZZЛ>+QJl޲9[-\5Nݻy@ym Jq`J ݤEBuR:6>{hge'&mI1+NAҲ5+n!HwzMC# trAx(<əY1 ڤAthO~2f38Z^`c3bBi1,.v`kt ^LTo+bAۊ(Ft(W}0>펉 x@ yRT* 9[%l71\Mxev$*,}& I(gM 7 &DRxxktB3Ąԏao+b޼\$^&ZX4ƼKrmr\, y F(x92A:^K-"I<=\6`%kϦ1pCJAI7 &I6/!Q\e ^Q,ĄGjWD G]z,$rT j44y4b X0j ef9FL~ JW¬݊_je c=\f EJVy\/K_O[J;8&dY־ۇIsd?̙ \Ɩ6f~.l&MzVAY%w׺FM$>ǃt% FyChZ#5N-i2UgoLU=Pk,uӍ:IK v*l{]*lΥZu*lƹ[nى~+7 ;<ן>_'`:Aq`} ZW' xk6 Zip݅ yB Dt0vgUk;y{:=Q=&؞j̶T~܃jySckLAj>iz@2hD+N ӷ&%hpXb23I c'y5ש KPog )iI&GVh^̳6$I\@DGzn@P 2.=5LEڛRɀfqr՚;tT҅ҭOLeT5*{|+Z؝U`3<~{ JfS^=Qo*%MDõQ 2&Q2z˖jAqnxt@ /ifҶj.E0i5gbY|j>Ġ3;wVA1!;Bmo_lH`j=Qh{f_H}1I.kϧH pgkXb%gM7{i7+q/qbPtv{N;k;n++/5*/|=oe [<[mYUb]ynQ xbG\|wOꜜ;tWMMz,ôX!|\yשElGg/ngwi*(Y+sYjۅ.ޣt^x?Hf'g6h7"Y?%Vg6LaϿSֻϲSgaH">۠L>;3^VL(RFqFy)7K,$E~~+wOuh ދm% /7vr8nCu؛QCĄzn&z2c5_"dʀh{&J~i>MDjVvFd]^۔O~LaJj~umtNw+D\/¯iq>"r 3l.}~k>M´Of`|Tb߹9}`޴~yJTBi'r78_j%_/>sVҠn0 (]c<(":46_)4[j5Z)LBbظ(u)|v22qSIkzߒ .7'f ٔm!g76`>abÕ,VZX&7T-x3pQD# &8wZtT)Qʁ WtFQ:,ZmJ(9s|wwҸY\9z$<s&q@ߚ+E^h:o]12\ T\xIXx-OC]b+i>EXeb1¿#KU|YM\(=|rzZwц GfRGX{qTFjd!+e!'jv鎨_]! 9)`[U!TG Zȯ{*K)ɐ lr1JL2XFx+% r&}IJtB TزkrO$ W{3C0$J?8QK'ZNT5j3U*<*+E48cCIYFF_Б/Bõi$}\_D5Q(Iv h vV G[uUݏݚr:z) V2/)S7Uז2/DKZ+`wYzP L./ ~:Ddt kd,ӼN/etIf.rgVzV#*rUw|߁`a}cz#ڒ`$KԳƀrF%aQ5[%\~Ơ1JCdh|XƁt(9%%>#ƺ,*J~Ag !⌚tbU var/home/core/zuul-output/logs/kubelet.log0000644000000000000000005002160415155400324017672 0ustar rootrootMar 15 00:06:45 crc systemd[1]: Starting Kubernetes Kubelet... Mar 15 00:06:45 crc restorecon[4688]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:45 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 15 00:06:46 crc restorecon[4688]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 15 00:06:47 crc kubenswrapper[4861]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.124719 4861 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130248 4861 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130285 4861 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130295 4861 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130304 4861 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130312 4861 feature_gate.go:330] unrecognized feature gate: Example Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130322 4861 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130332 4861 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130340 4861 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130350 4861 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130358 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130367 4861 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130376 4861 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130384 4861 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130400 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130412 4861 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130424 4861 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130434 4861 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130443 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130452 4861 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130460 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130469 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130478 4861 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130488 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130497 4861 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130508 4861 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130518 4861 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130528 4861 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130537 4861 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130546 4861 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130581 4861 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130592 4861 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130603 4861 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130613 4861 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130623 4861 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130634 4861 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130645 4861 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130655 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130665 4861 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130674 4861 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130682 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130692 4861 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130700 4861 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130709 4861 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130719 4861 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130728 4861 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130736 4861 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130745 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130754 4861 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130763 4861 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130772 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130780 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130788 4861 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130800 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130809 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130817 4861 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130826 4861 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130834 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130842 4861 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130851 4861 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130859 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130867 4861 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130875 4861 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130883 4861 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130892 4861 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130900 4861 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130908 4861 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130918 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130941 4861 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130950 4861 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130958 4861 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.130967 4861 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.131926 4861 flags.go:64] FLAG: --address="0.0.0.0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.131953 4861 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.131968 4861 flags.go:64] FLAG: --anonymous-auth="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.131981 4861 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132000 4861 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132011 4861 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132024 4861 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132035 4861 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132046 4861 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132056 4861 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132068 4861 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132080 4861 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132092 4861 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132103 4861 flags.go:64] FLAG: --cgroup-root="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132112 4861 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132122 4861 flags.go:64] FLAG: --client-ca-file="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132131 4861 flags.go:64] FLAG: --cloud-config="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132141 4861 flags.go:64] FLAG: --cloud-provider="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132150 4861 flags.go:64] FLAG: --cluster-dns="[]" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132171 4861 flags.go:64] FLAG: --cluster-domain="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132180 4861 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132191 4861 flags.go:64] FLAG: --config-dir="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132200 4861 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132210 4861 flags.go:64] FLAG: --container-log-max-files="5" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132222 4861 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132232 4861 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132241 4861 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132252 4861 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132270 4861 flags.go:64] FLAG: --contention-profiling="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132280 4861 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132290 4861 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132301 4861 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132323 4861 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132335 4861 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132345 4861 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132354 4861 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132364 4861 flags.go:64] FLAG: --enable-load-reader="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132375 4861 flags.go:64] FLAG: --enable-server="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132384 4861 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132403 4861 flags.go:64] FLAG: --event-burst="100" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132413 4861 flags.go:64] FLAG: --event-qps="50" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132422 4861 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132432 4861 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132441 4861 flags.go:64] FLAG: --eviction-hard="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132455 4861 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132465 4861 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132475 4861 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132485 4861 flags.go:64] FLAG: --eviction-soft="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132494 4861 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132504 4861 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132514 4861 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132523 4861 flags.go:64] FLAG: --experimental-mounter-path="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132533 4861 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132542 4861 flags.go:64] FLAG: --fail-swap-on="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132552 4861 flags.go:64] FLAG: --feature-gates="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132602 4861 flags.go:64] FLAG: --file-check-frequency="20s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132615 4861 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132628 4861 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132640 4861 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132653 4861 flags.go:64] FLAG: --healthz-port="10248" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132666 4861 flags.go:64] FLAG: --help="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132676 4861 flags.go:64] FLAG: --hostname-override="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132685 4861 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132695 4861 flags.go:64] FLAG: --http-check-frequency="20s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132704 4861 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132714 4861 flags.go:64] FLAG: --image-credential-provider-config="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132725 4861 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132734 4861 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132767 4861 flags.go:64] FLAG: --image-service-endpoint="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132777 4861 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132786 4861 flags.go:64] FLAG: --kube-api-burst="100" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132797 4861 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132807 4861 flags.go:64] FLAG: --kube-api-qps="50" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132816 4861 flags.go:64] FLAG: --kube-reserved="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132826 4861 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132835 4861 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132847 4861 flags.go:64] FLAG: --kubelet-cgroups="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132858 4861 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132871 4861 flags.go:64] FLAG: --lock-file="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132883 4861 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132896 4861 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132911 4861 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132932 4861 flags.go:64] FLAG: --log-json-split-stream="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132945 4861 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132958 4861 flags.go:64] FLAG: --log-text-split-stream="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132967 4861 flags.go:64] FLAG: --logging-format="text" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132977 4861 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132988 4861 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.132997 4861 flags.go:64] FLAG: --manifest-url="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133007 4861 flags.go:64] FLAG: --manifest-url-header="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133019 4861 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133029 4861 flags.go:64] FLAG: --max-open-files="1000000" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133041 4861 flags.go:64] FLAG: --max-pods="110" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133051 4861 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133062 4861 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133072 4861 flags.go:64] FLAG: --memory-manager-policy="None" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133081 4861 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133091 4861 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133101 4861 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133110 4861 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133131 4861 flags.go:64] FLAG: --node-status-max-images="50" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133140 4861 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133150 4861 flags.go:64] FLAG: --oom-score-adj="-999" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133160 4861 flags.go:64] FLAG: --pod-cidr="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133182 4861 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133196 4861 flags.go:64] FLAG: --pod-manifest-path="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133218 4861 flags.go:64] FLAG: --pod-max-pids="-1" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133229 4861 flags.go:64] FLAG: --pods-per-core="0" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133239 4861 flags.go:64] FLAG: --port="10250" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133249 4861 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133258 4861 flags.go:64] FLAG: --provider-id="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133268 4861 flags.go:64] FLAG: --qos-reserved="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133277 4861 flags.go:64] FLAG: --read-only-port="10255" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133287 4861 flags.go:64] FLAG: --register-node="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133296 4861 flags.go:64] FLAG: --register-schedulable="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133306 4861 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133324 4861 flags.go:64] FLAG: --registry-burst="10" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133334 4861 flags.go:64] FLAG: --registry-qps="5" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133343 4861 flags.go:64] FLAG: --reserved-cpus="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133352 4861 flags.go:64] FLAG: --reserved-memory="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133364 4861 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133374 4861 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133384 4861 flags.go:64] FLAG: --rotate-certificates="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133393 4861 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133412 4861 flags.go:64] FLAG: --runonce="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133422 4861 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133431 4861 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133441 4861 flags.go:64] FLAG: --seccomp-default="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133451 4861 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133460 4861 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133470 4861 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133480 4861 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133490 4861 flags.go:64] FLAG: --storage-driver-password="root" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133499 4861 flags.go:64] FLAG: --storage-driver-secure="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133509 4861 flags.go:64] FLAG: --storage-driver-table="stats" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133518 4861 flags.go:64] FLAG: --storage-driver-user="root" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133528 4861 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133538 4861 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133548 4861 flags.go:64] FLAG: --system-cgroups="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133593 4861 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133624 4861 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133634 4861 flags.go:64] FLAG: --tls-cert-file="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133643 4861 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133662 4861 flags.go:64] FLAG: --tls-min-version="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133672 4861 flags.go:64] FLAG: --tls-private-key-file="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133681 4861 flags.go:64] FLAG: --topology-manager-policy="none" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133691 4861 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133700 4861 flags.go:64] FLAG: --topology-manager-scope="container" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133710 4861 flags.go:64] FLAG: --v="2" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133723 4861 flags.go:64] FLAG: --version="false" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133735 4861 flags.go:64] FLAG: --vmodule="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133746 4861 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.133756 4861 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134029 4861 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134049 4861 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134062 4861 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134077 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134087 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134096 4861 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134106 4861 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134117 4861 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134127 4861 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134137 4861 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134146 4861 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134156 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134165 4861 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134174 4861 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134183 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134192 4861 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134200 4861 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134209 4861 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134218 4861 feature_gate.go:330] unrecognized feature gate: Example Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134226 4861 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134237 4861 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134247 4861 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134257 4861 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134268 4861 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134278 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134289 4861 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134299 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134309 4861 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134320 4861 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134331 4861 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134341 4861 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134351 4861 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134361 4861 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134376 4861 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134386 4861 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134401 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134411 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134420 4861 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134429 4861 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134437 4861 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134445 4861 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134454 4861 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134462 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134470 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134480 4861 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134490 4861 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134500 4861 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134510 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134521 4861 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134530 4861 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134542 4861 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134553 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134596 4861 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134606 4861 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134617 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134627 4861 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134637 4861 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134645 4861 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134654 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134666 4861 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134675 4861 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134683 4861 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134691 4861 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134700 4861 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134708 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134724 4861 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134735 4861 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134749 4861 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134758 4861 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134768 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.134778 4861 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.134809 4861 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.149824 4861 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.150187 4861 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150331 4861 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150347 4861 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150357 4861 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150368 4861 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150379 4861 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150389 4861 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150401 4861 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150412 4861 feature_gate.go:330] unrecognized feature gate: Example Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150422 4861 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150435 4861 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150446 4861 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150457 4861 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150470 4861 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150481 4861 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150493 4861 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150504 4861 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150515 4861 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150523 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150531 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150540 4861 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150549 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150586 4861 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150594 4861 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150604 4861 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150612 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150621 4861 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150630 4861 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150642 4861 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150654 4861 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150663 4861 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150675 4861 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150690 4861 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150701 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150711 4861 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150729 4861 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150752 4861 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150766 4861 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150777 4861 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150787 4861 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150798 4861 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150809 4861 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150820 4861 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150831 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150842 4861 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150853 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150864 4861 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150876 4861 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150887 4861 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150898 4861 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150908 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150918 4861 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150929 4861 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150940 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150948 4861 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150957 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150965 4861 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150976 4861 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.150987 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151000 4861 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151015 4861 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151029 4861 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151042 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151057 4861 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151070 4861 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151081 4861 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151091 4861 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151137 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151148 4861 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151157 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151166 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151177 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.151192 4861 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151493 4861 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151520 4861 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151533 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151544 4861 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151592 4861 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151603 4861 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151614 4861 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151624 4861 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151634 4861 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151642 4861 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151654 4861 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151663 4861 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151671 4861 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151680 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151689 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151697 4861 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151706 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151714 4861 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151722 4861 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151735 4861 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151745 4861 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151754 4861 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151763 4861 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151775 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151786 4861 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151797 4861 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151809 4861 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151820 4861 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151831 4861 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151841 4861 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151854 4861 feature_gate.go:330] unrecognized feature gate: Example Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151865 4861 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151876 4861 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151886 4861 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151899 4861 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151909 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151918 4861 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151927 4861 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151935 4861 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151944 4861 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151952 4861 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151963 4861 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151975 4861 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151987 4861 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.151999 4861 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152011 4861 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152023 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152035 4861 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152045 4861 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152056 4861 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152067 4861 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152080 4861 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152094 4861 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152105 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152116 4861 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152130 4861 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152141 4861 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152153 4861 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152164 4861 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152174 4861 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152184 4861 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152195 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152205 4861 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152217 4861 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152227 4861 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152237 4861 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152248 4861 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152259 4861 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152270 4861 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152280 4861 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.152293 4861 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.152310 4861 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.153762 4861 server.go:940] "Client rotation is on, will bootstrap in background" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.159417 4861 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.165286 4861 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.165458 4861 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.167068 4861 server.go:997] "Starting client certificate rotation" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.167120 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.167306 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.198012 4861 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.199673 4861 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.203614 4861 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.223903 4861 log.go:25] "Validated CRI v1 runtime API" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.266221 4861 log.go:25] "Validated CRI v1 image API" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.269101 4861 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.274538 4861 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-15-00-01-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.274635 4861 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.302446 4861 manager.go:217] Machine: {Timestamp:2026-03-15 00:06:47.299709258 +0000 UTC m=+0.611906821 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0787823d-1679-41ae-95f7-6ed423e90e04 BootID:963bfba1-4eec-4efc-bcce-77026d8b2488 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e9:9e:b7 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e9:9e:b7 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:98:33:e8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:ba:09:d0 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9d:8e:60 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:8e:aa:95 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4e:e1:5d:45:39:8a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:46:71:e9:a5:b7:b9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.302947 4861 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.303299 4861 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.303942 4861 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.304253 4861 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.304308 4861 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.304690 4861 topology_manager.go:138] "Creating topology manager with none policy" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.304710 4861 container_manager_linux.go:303] "Creating device plugin manager" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.305162 4861 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.305211 4861 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.305534 4861 state_mem.go:36] "Initialized new in-memory state store" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.305748 4861 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.309239 4861 kubelet.go:418] "Attempting to sync node with API server" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.309494 4861 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.309594 4861 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.309624 4861 kubelet.go:324] "Adding apiserver pod source" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.309647 4861 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.315014 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.317183 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.317693 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.317820 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.320065 4861 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.321364 4861 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.324282 4861 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326619 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326670 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326688 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326702 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326726 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326740 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326755 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326779 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326796 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326812 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326852 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.326867 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.328028 4861 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.328765 4861 server.go:1280] "Started kubelet" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.329269 4861 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.329692 4861 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.330189 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.330828 4861 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 15 00:06:47 crc systemd[1]: Started Kubernetes Kubelet. Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.333743 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.333801 4861 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.334436 4861 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.334653 4861 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.334980 4861 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.334457 4861 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.335252 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.335633 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.335220 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.335976 4861 factory.go:55] Registering systemd factory Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.335998 4861 factory.go:221] Registration of the systemd container factory successfully Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336029 4861 server.go:460] "Adding debug handlers to kubelet server" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336533 4861 factory.go:153] Registering CRI-O factory Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336551 4861 factory.go:221] Registration of the crio container factory successfully Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336623 4861 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336638 4861 factory.go:103] Registering Raw factory Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.336650 4861 manager.go:1196] Started watching for new ooms in manager Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.337227 4861 manager.go:319] Starting recovery of all containers Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.340335 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365614 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365662 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365676 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365687 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365699 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365709 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365719 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365729 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365744 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365756 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365767 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365777 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365788 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365800 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365811 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365823 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365835 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365844 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365856 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365867 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365876 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365899 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365910 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365922 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365934 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365947 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365963 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365977 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365987 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.365997 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366009 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366021 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366034 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366045 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366054 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366063 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366075 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366084 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366097 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366109 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366120 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366132 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366142 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366153 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366165 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366176 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366186 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366196 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366207 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366219 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366231 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366248 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366266 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366306 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366322 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366333 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366346 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366357 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366368 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366378 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366389 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366399 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366412 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366425 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366437 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366448 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366461 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366473 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366484 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366495 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366508 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366520 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366533 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366546 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366573 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366584 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366594 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366605 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366616 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366627 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366639 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366651 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366665 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366676 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366693 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366705 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366731 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366742 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366755 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366766 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366776 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366788 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366799 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366810 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366820 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366831 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366840 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366850 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366862 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366872 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366883 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366899 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366913 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366925 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366942 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366954 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366965 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.366977 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367010 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367023 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367034 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367047 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367060 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367071 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367084 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367095 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367106 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367118 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367129 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367141 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367151 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367161 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367173 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367185 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367196 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367206 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367274 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367285 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367295 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367306 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367318 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367330 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367341 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367352 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367365 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367376 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367388 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367401 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367412 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367422 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367433 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367446 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367456 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367466 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367478 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367490 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367505 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367517 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367527 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367540 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367552 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367577 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367588 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367600 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367611 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367626 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367636 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367648 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367659 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367670 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367681 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367692 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367704 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367715 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367726 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367736 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367750 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367760 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367769 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367778 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367788 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367798 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367808 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367820 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367831 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367843 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367853 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.367864 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369776 4861 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369800 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369811 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369819 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369830 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369841 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369852 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369864 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369879 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369891 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369901 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369912 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369922 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369935 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369948 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369958 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369970 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369979 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369988 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.369999 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370010 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370019 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370028 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370038 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370049 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370058 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370066 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370076 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370087 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370097 4861 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370105 4861 reconstruct.go:97] "Volume reconstruction finished" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.370113 4861 reconciler.go:26] "Reconciler: start to sync state" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.374813 4861 manager.go:324] Recovery completed Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.389202 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.392921 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.393006 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.393027 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.395138 4861 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.395196 4861 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.395224 4861 state_mem.go:36] "Initialized new in-memory state store" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.405589 4861 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.407527 4861 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.407590 4861 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.407628 4861 kubelet.go:2335] "Starting kubelet main sync loop" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.407685 4861 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.408947 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.409029 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.411908 4861 policy_none.go:49] "None policy: Start" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.413211 4861 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.413355 4861 state_mem.go:35] "Initializing new in-memory state store" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.435612 4861 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.466910 4861 manager.go:334] "Starting Device Plugin manager" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.467009 4861 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.467026 4861 server.go:79] "Starting device plugin registration server" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.467670 4861 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.467697 4861 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.467904 4861 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.468005 4861 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.468014 4861 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.474163 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.508484 4861 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.508687 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.509893 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.509935 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.509948 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510118 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510660 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510769 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510776 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510802 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510819 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.510981 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.511194 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.511256 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512215 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512251 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512267 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512389 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512461 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512471 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512641 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512597 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.512745 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513137 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513168 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513185 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513330 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513361 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513386 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513424 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513460 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513477 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513714 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513798 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.513831 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.514890 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.514941 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.514958 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.515025 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.515046 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.515059 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.515354 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.515382 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.516048 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.516106 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.516119 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.536474 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.546526 4861 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/cpuset.cpus.effective": open /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/cpuset.cpus.effective: no such device Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.568077 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.569930 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.569990 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.570008 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.570043 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.570732 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571537 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571590 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571621 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571673 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571694 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571719 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571737 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571751 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571770 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571831 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571872 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571892 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571913 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571930 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.571966 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673204 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673270 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673298 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673321 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673344 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673364 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673386 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673407 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673432 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673423 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673460 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673484 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673498 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673542 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673585 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673591 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673596 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673617 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673619 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673542 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673609 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673637 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673667 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673641 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673677 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673512 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673502 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673661 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673551 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.673618 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.771093 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.772765 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.772838 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.772857 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.772894 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.773639 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.844621 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.863506 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.871796 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.884866 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: I0315 00:06:47.890807 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.897285 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-bbca0c7caec2bab13f5a7001ccbedf2564c563a7e321797d679ede703a31939e WatchSource:0}: Error finding container bbca0c7caec2bab13f5a7001ccbedf2564c563a7e321797d679ede703a31939e: Status 404 returned error can't find the container with id bbca0c7caec2bab13f5a7001ccbedf2564c563a7e321797d679ede703a31939e Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.908510 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4fbd65b6db372be7ae000985e9fffd0b2948e2925b2fc9bf642b9ce9dad9bb7b WatchSource:0}: Error finding container 4fbd65b6db372be7ae000985e9fffd0b2948e2925b2fc9bf642b9ce9dad9bb7b: Status 404 returned error can't find the container with id 4fbd65b6db372be7ae000985e9fffd0b2948e2925b2fc9bf642b9ce9dad9bb7b Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.912997 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-dd0b29454f439ca336c9fa39046b0b62d5b37284b09dff9703f10e50fa15187e WatchSource:0}: Error finding container dd0b29454f439ca336c9fa39046b0b62d5b37284b09dff9703f10e50fa15187e: Status 404 returned error can't find the container with id dd0b29454f439ca336c9fa39046b0b62d5b37284b09dff9703f10e50fa15187e Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.925722 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-bcaf3ee33ef479e1c7a1ef6c1f7fb29f895ad7ccd41b449b717f7dc5f982d8f1 WatchSource:0}: Error finding container bcaf3ee33ef479e1c7a1ef6c1f7fb29f895ad7ccd41b449b717f7dc5f982d8f1: Status 404 returned error can't find the container with id bcaf3ee33ef479e1c7a1ef6c1f7fb29f895ad7ccd41b449b717f7dc5f982d8f1 Mar 15 00:06:47 crc kubenswrapper[4861]: W0315 00:06:47.929098 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fd144a1c474768f251b7e2b4cf294542f687fa9f3a9baf1a468d21720f29d827 WatchSource:0}: Error finding container fd144a1c474768f251b7e2b4cf294542f687fa9f3a9baf1a468d21720f29d827: Status 404 returned error can't find the container with id fd144a1c474768f251b7e2b4cf294542f687fa9f3a9baf1a468d21720f29d827 Mar 15 00:06:47 crc kubenswrapper[4861]: E0315 00:06:47.937784 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.174546 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.176349 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.176392 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.176408 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.176437 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.176927 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.331065 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:48 crc kubenswrapper[4861]: W0315 00:06:48.399533 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.399672 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.412315 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bbca0c7caec2bab13f5a7001ccbedf2564c563a7e321797d679ede703a31939e"} Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.413229 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fd144a1c474768f251b7e2b4cf294542f687fa9f3a9baf1a468d21720f29d827"} Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.414399 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bcaf3ee33ef479e1c7a1ef6c1f7fb29f895ad7ccd41b449b717f7dc5f982d8f1"} Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.416290 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dd0b29454f439ca336c9fa39046b0b62d5b37284b09dff9703f10e50fa15187e"} Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.417549 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4fbd65b6db372be7ae000985e9fffd0b2948e2925b2fc9bf642b9ce9dad9bb7b"} Mar 15 00:06:48 crc kubenswrapper[4861]: W0315 00:06:48.419227 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.419296 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:48 crc kubenswrapper[4861]: W0315 00:06:48.545883 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.545959 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.739396 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 15 00:06:48 crc kubenswrapper[4861]: W0315 00:06:48.902469 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.902649 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.977309 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.979409 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.979440 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.979449 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:48 crc kubenswrapper[4861]: I0315 00:06:48.979468 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:48 crc kubenswrapper[4861]: E0315 00:06:48.980010 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.331367 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.394104 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:06:49 crc kubenswrapper[4861]: E0315 00:06:49.395946 4861 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.423929 4861 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605" exitCode=0 Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.424010 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.424109 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.425389 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.425426 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.425443 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.427263 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964" exitCode=0 Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.427306 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.427346 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.428383 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.428425 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.428437 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.430417 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.430699 4861 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55" exitCode=0 Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.430796 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.430823 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.431335 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.431383 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.431396 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.431986 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.432017 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.432027 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.433141 4861 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611" exitCode=0 Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.433325 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.433961 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.434486 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.434507 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.434517 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.441875 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.441948 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.441984 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.442001 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968"} Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.441954 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.443198 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.443240 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.443253 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:49 crc kubenswrapper[4861]: I0315 00:06:49.640933 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:50 crc kubenswrapper[4861]: W0315 00:06:50.275152 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:50 crc kubenswrapper[4861]: E0315 00:06:50.275250 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.333504 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 15 00:06:50 crc kubenswrapper[4861]: E0315 00:06:50.341021 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.448920 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.449051 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.450272 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.450304 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.450314 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.451765 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.451833 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.451857 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.451908 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.453287 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.453322 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.453332 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.454548 4861 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b" exitCode=0 Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.454681 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.454707 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.455447 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.455469 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.455494 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.458593 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.458628 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.458640 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.458650 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667"} Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.458673 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.459407 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.459433 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.459445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.580292 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.581903 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.581976 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.581989 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.582023 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:50 crc kubenswrapper[4861]: E0315 00:06:50.582655 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 15 00:06:50 crc kubenswrapper[4861]: I0315 00:06:50.818518 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.464359 4861 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039" exitCode=0 Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.464445 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039"} Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.464649 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.465936 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.465993 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.466018 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.468624 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.468776 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.468619 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4acbb519765c8c69d574598be9f572d46d6bd220dab1fc4879eb00193cdd8beb"} Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.469016 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.468795 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.469114 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.469314 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.469348 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.469371 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470237 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470277 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470287 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470456 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470492 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470514 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470587 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470605 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.470616 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:51 crc kubenswrapper[4861]: I0315 00:06:51.656746 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478439 4861 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478475 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01"} Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478601 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478589 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc"} Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478778 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5"} Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478617 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.478506 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480420 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480449 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480459 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480629 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480678 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480721 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480739 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480687 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:52 crc kubenswrapper[4861]: I0315 00:06:52.480801 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.469642 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.489056 4861 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.489035 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca"} Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.489142 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed"} Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.489150 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.489215 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491127 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491176 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491504 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491651 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.491691 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.783841 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.786397 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.786459 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.786478 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.786513 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.819240 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:06:53 crc kubenswrapper[4861]: I0315 00:06:53.819311 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.278422 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.392911 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.491518 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.491613 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493185 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493252 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493201 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493279 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:54 crc kubenswrapper[4861]: I0315 00:06:54.493427 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.495009 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.496610 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.496674 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.496692 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.538375 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.538811 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.540423 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.540478 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.540502 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.777188 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.777515 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.779457 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.779553 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:55 crc kubenswrapper[4861]: I0315 00:06:55.779617 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:56 crc kubenswrapper[4861]: I0315 00:06:56.033693 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 15 00:06:56 crc kubenswrapper[4861]: I0315 00:06:56.499005 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:56 crc kubenswrapper[4861]: I0315 00:06:56.500452 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:56 crc kubenswrapper[4861]: I0315 00:06:56.500529 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:56 crc kubenswrapper[4861]: I0315 00:06:56.500549 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.003961 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.004267 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.006431 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.006501 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.006532 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.016466 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:57 crc kubenswrapper[4861]: E0315 00:06:57.474283 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.502546 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.504787 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.504869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:57 crc kubenswrapper[4861]: I0315 00:06:57.504891 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:06:59 crc kubenswrapper[4861]: I0315 00:06:59.646012 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:06:59 crc kubenswrapper[4861]: I0315 00:06:59.646181 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:06:59 crc kubenswrapper[4861]: I0315 00:06:59.647706 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:06:59 crc kubenswrapper[4861]: I0315 00:06:59.647752 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:06:59 crc kubenswrapper[4861]: I0315 00:06:59.647765 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:01 crc kubenswrapper[4861]: W0315 00:07:01.183684 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.183869 4861 trace.go:236] Trace[1815619646]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Mar-2026 00:06:51.181) (total time: 10002ms): Mar 15 00:07:01 crc kubenswrapper[4861]: Trace[1815619646]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:07:01.183) Mar 15 00:07:01 crc kubenswrapper[4861]: Trace[1815619646]: [10.002116752s] [10.002116752s] END Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.183912 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.332782 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 15 00:07:01 crc kubenswrapper[4861]: W0315 00:07:01.439279 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.439359 4861 trace.go:236] Trace[1578223099]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (15-Mar-2026 00:06:51.438) (total time: 10000ms): Mar 15 00:07:01 crc kubenswrapper[4861]: Trace[1578223099]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (00:07:01.439) Mar 15 00:07:01 crc kubenswrapper[4861]: Trace[1578223099]: [10.000971841s] [10.000971841s] END Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.439378 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.593988 4861 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:01 crc kubenswrapper[4861]: W0315 00:07:01.595499 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.595624 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.597833 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.598762 4861 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.598833 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.605863 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.606817 4861 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.606876 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.614081 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:01 crc kubenswrapper[4861]: W0315 00:07:01.614984 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z Mar 15 00:07:01 crc kubenswrapper[4861]: E0315 00:07:01.615123 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.662595 4861 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:01 crc kubenswrapper[4861]: I0315 00:07:01.662657 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.335290 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:02Z is after 2026-02-23T05:33:13Z Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.521109 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.523989 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4acbb519765c8c69d574598be9f572d46d6bd220dab1fc4879eb00193cdd8beb" exitCode=255 Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.524067 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4acbb519765c8c69d574598be9f572d46d6bd220dab1fc4879eb00193cdd8beb"} Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.524425 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.525910 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.525966 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.525981 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:02 crc kubenswrapper[4861]: I0315 00:07:02.527053 4861 scope.go:117] "RemoveContainer" containerID="4acbb519765c8c69d574598be9f572d46d6bd220dab1fc4879eb00193cdd8beb" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.335471 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:03Z is after 2026-02-23T05:33:13Z Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.528258 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.530761 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520"} Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.530911 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.531693 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.531730 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.531743 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.819260 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:07:03 crc kubenswrapper[4861]: I0315 00:07:03.819406 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.278523 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.335843 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:04Z is after 2026-02-23T05:33:13Z Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.536487 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.537124 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.540188 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" exitCode=255 Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.540229 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520"} Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.540283 4861 scope.go:117] "RemoveContainer" containerID="4acbb519765c8c69d574598be9f572d46d6bd220dab1fc4879eb00193cdd8beb" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.540341 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.541746 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.541780 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.541790 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:04 crc kubenswrapper[4861]: I0315 00:07:04.542293 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:04 crc kubenswrapper[4861]: E0315 00:07:04.542466 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:05 crc kubenswrapper[4861]: W0315 00:07:05.145729 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z Mar 15 00:07:05 crc kubenswrapper[4861]: E0315 00:07:05.146183 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:05 crc kubenswrapper[4861]: W0315 00:07:05.158672 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z Mar 15 00:07:05 crc kubenswrapper[4861]: E0315 00:07:05.158754 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.334327 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.546443 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.550191 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.551669 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.551727 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.551746 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:05 crc kubenswrapper[4861]: I0315 00:07:05.552856 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:05 crc kubenswrapper[4861]: E0315 00:07:05.553168 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:05 crc kubenswrapper[4861]: W0315 00:07:05.686001 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z Mar 15 00:07:05 crc kubenswrapper[4861]: E0315 00:07:05.686152 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.071940 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.072344 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.074697 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.074773 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.074791 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.093473 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.336019 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:06Z is after 2026-02-23T05:33:13Z Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.552777 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.553884 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.553926 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.553936 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.664037 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.664273 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.665535 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.665612 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.665627 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.666462 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:06 crc kubenswrapper[4861]: E0315 00:07:06.666702 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:06 crc kubenswrapper[4861]: I0315 00:07:06.671454 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.337688 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:07Z is after 2026-02-23T05:33:13Z Mar 15 00:07:07 crc kubenswrapper[4861]: E0315 00:07:07.474427 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.516816 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.555400 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.556474 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.556532 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.556553 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:07 crc kubenswrapper[4861]: I0315 00:07:07.557668 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:07 crc kubenswrapper[4861]: E0315 00:07:07.558102 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:08 crc kubenswrapper[4861]: E0315 00:07:08.002311 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:08Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.006265 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.008046 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.008093 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.008112 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.008156 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:08 crc kubenswrapper[4861]: E0315 00:07:08.012149 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:08Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.336502 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:08Z is after 2026-02-23T05:33:13Z Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.558097 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.559336 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.559405 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.559431 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:08 crc kubenswrapper[4861]: I0315 00:07:08.560389 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:08 crc kubenswrapper[4861]: E0315 00:07:08.560776 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:09 crc kubenswrapper[4861]: I0315 00:07:09.335068 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:09Z is after 2026-02-23T05:33:13Z Mar 15 00:07:09 crc kubenswrapper[4861]: I0315 00:07:09.695337 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:07:09 crc kubenswrapper[4861]: E0315 00:07:09.699677 4861 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:09Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:10 crc kubenswrapper[4861]: I0315 00:07:10.336002 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:10Z is after 2026-02-23T05:33:13Z Mar 15 00:07:11 crc kubenswrapper[4861]: I0315 00:07:11.334317 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:11Z is after 2026-02-23T05:33:13Z Mar 15 00:07:11 crc kubenswrapper[4861]: E0315 00:07:11.619838 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:11Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:12 crc kubenswrapper[4861]: I0315 00:07:12.336932 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:12Z is after 2026-02-23T05:33:13Z Mar 15 00:07:12 crc kubenswrapper[4861]: W0315 00:07:12.769688 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:12Z is after 2026-02-23T05:33:13Z Mar 15 00:07:12 crc kubenswrapper[4861]: E0315 00:07:12.769828 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:12Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.337964 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:13Z is after 2026-02-23T05:33:13Z Mar 15 00:07:13 crc kubenswrapper[4861]: W0315 00:07:13.809472 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:13Z is after 2026-02-23T05:33:13Z Mar 15 00:07:13 crc kubenswrapper[4861]: E0315 00:07:13.809652 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:13Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.819511 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.819665 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.819787 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.820058 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.822280 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.822521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.822775 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.823955 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 15 00:07:13 crc kubenswrapper[4861]: I0315 00:07:13.824397 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28" gracePeriod=30 Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.337482 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:14Z is after 2026-02-23T05:33:13Z Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.581385 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.583373 4861 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28" exitCode=255 Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.583454 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28"} Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.583512 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb"} Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.583704 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.585482 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.585553 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:14 crc kubenswrapper[4861]: I0315 00:07:14.585616 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:15 crc kubenswrapper[4861]: E0315 00:07:15.008109 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:15Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.013199 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.015106 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.015206 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.015232 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.015290 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:15 crc kubenswrapper[4861]: E0315 00:07:15.021744 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:15Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.336656 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:15Z is after 2026-02-23T05:33:13Z Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.778066 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.778354 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.780386 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.780442 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:15 crc kubenswrapper[4861]: I0315 00:07:15.780466 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:16 crc kubenswrapper[4861]: I0315 00:07:16.335534 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:16Z is after 2026-02-23T05:33:13Z Mar 15 00:07:16 crc kubenswrapper[4861]: W0315 00:07:16.768373 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:16Z is after 2026-02-23T05:33:13Z Mar 15 00:07:16 crc kubenswrapper[4861]: E0315 00:07:16.768530 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:17 crc kubenswrapper[4861]: W0315 00:07:17.129944 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:17Z is after 2026-02-23T05:33:13Z Mar 15 00:07:17 crc kubenswrapper[4861]: E0315 00:07:17.130087 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:17 crc kubenswrapper[4861]: I0315 00:07:17.336548 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:17Z is after 2026-02-23T05:33:13Z Mar 15 00:07:17 crc kubenswrapper[4861]: E0315 00:07:17.474648 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:18 crc kubenswrapper[4861]: I0315 00:07:18.334887 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:18Z is after 2026-02-23T05:33:13Z Mar 15 00:07:19 crc kubenswrapper[4861]: I0315 00:07:19.333534 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:19Z is after 2026-02-23T05:33:13Z Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.334445 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:20Z is after 2026-02-23T05:33:13Z Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.819532 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.819787 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.821219 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.821267 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:20 crc kubenswrapper[4861]: I0315 00:07:20.821285 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.333867 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:21Z is after 2026-02-23T05:33:13Z Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.408843 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.411002 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.411073 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.411094 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:21 crc kubenswrapper[4861]: I0315 00:07:21.412187 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:21 crc kubenswrapper[4861]: E0315 00:07:21.626058 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:21Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:22 crc kubenswrapper[4861]: E0315 00:07:22.012421 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.022672 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.023877 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.023903 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.023912 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.023932 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:22 crc kubenswrapper[4861]: E0315 00:07:22.026874 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.334199 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:22Z is after 2026-02-23T05:33:13Z Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.611394 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.612648 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.615591 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" exitCode=255 Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.615666 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad"} Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.615730 4861 scope.go:117] "RemoveContainer" containerID="21b5b7ee2bd3bfea1bd01c7161aceedc7c0c75fb9e205a8f0383c3ea59a9f520" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.615940 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.617432 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.617481 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.617759 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:22 crc kubenswrapper[4861]: I0315 00:07:22.618726 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:22 crc kubenswrapper[4861]: E0315 00:07:22.619042 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:23 crc kubenswrapper[4861]: I0315 00:07:23.336865 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:23Z is after 2026-02-23T05:33:13Z Mar 15 00:07:23 crc kubenswrapper[4861]: I0315 00:07:23.620060 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 15 00:07:23 crc kubenswrapper[4861]: I0315 00:07:23.819742 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:07:23 crc kubenswrapper[4861]: I0315 00:07:23.819823 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.279216 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.279434 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.280853 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.280922 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.280939 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.281775 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:24 crc kubenswrapper[4861]: E0315 00:07:24.282137 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:24 crc kubenswrapper[4861]: I0315 00:07:24.334473 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:24Z is after 2026-02-23T05:33:13Z Mar 15 00:07:25 crc kubenswrapper[4861]: I0315 00:07:25.334954 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:25Z is after 2026-02-23T05:33:13Z Mar 15 00:07:25 crc kubenswrapper[4861]: I0315 00:07:25.719923 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:07:25 crc kubenswrapper[4861]: E0315 00:07:25.726082 4861 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:25 crc kubenswrapper[4861]: E0315 00:07:25.727354 4861 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 15 00:07:26 crc kubenswrapper[4861]: I0315 00:07:26.337031 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:26Z is after 2026-02-23T05:33:13Z Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.333668 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:27Z is after 2026-02-23T05:33:13Z Mar 15 00:07:27 crc kubenswrapper[4861]: E0315 00:07:27.474761 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.516822 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.517011 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.517939 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.517998 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.518010 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:27 crc kubenswrapper[4861]: I0315 00:07:27.518642 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:27 crc kubenswrapper[4861]: E0315 00:07:27.518841 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:28 crc kubenswrapper[4861]: I0315 00:07:28.334791 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:28Z is after 2026-02-23T05:33:13Z Mar 15 00:07:28 crc kubenswrapper[4861]: W0315 00:07:28.682875 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:28Z is after 2026-02-23T05:33:13Z Mar 15 00:07:28 crc kubenswrapper[4861]: E0315 00:07:28.682977 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:29 crc kubenswrapper[4861]: E0315 00:07:29.018026 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.027365 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.028783 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.028850 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.028874 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.028916 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:29 crc kubenswrapper[4861]: E0315 00:07:29.033767 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:29 crc kubenswrapper[4861]: I0315 00:07:29.338493 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:29Z is after 2026-02-23T05:33:13Z Mar 15 00:07:30 crc kubenswrapper[4861]: I0315 00:07:30.335086 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:30Z is after 2026-02-23T05:33:13Z Mar 15 00:07:30 crc kubenswrapper[4861]: W0315 00:07:30.937960 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:30Z is after 2026-02-23T05:33:13Z Mar 15 00:07:30 crc kubenswrapper[4861]: E0315 00:07:30.938061 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:30Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:31 crc kubenswrapper[4861]: I0315 00:07:31.336731 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:31Z is after 2026-02-23T05:33:13Z Mar 15 00:07:31 crc kubenswrapper[4861]: E0315 00:07:31.630531 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:31Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:32 crc kubenswrapper[4861]: I0315 00:07:32.336154 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:32Z is after 2026-02-23T05:33:13Z Mar 15 00:07:33 crc kubenswrapper[4861]: I0315 00:07:33.333415 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:33Z is after 2026-02-23T05:33:13Z Mar 15 00:07:33 crc kubenswrapper[4861]: I0315 00:07:33.819866 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:07:33 crc kubenswrapper[4861]: I0315 00:07:33.819972 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:07:34 crc kubenswrapper[4861]: I0315 00:07:34.334282 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:34Z is after 2026-02-23T05:33:13Z Mar 15 00:07:35 crc kubenswrapper[4861]: I0315 00:07:35.334681 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:35Z is after 2026-02-23T05:33:13Z Mar 15 00:07:36 crc kubenswrapper[4861]: E0315 00:07:36.021490 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:36Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.034862 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.036100 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.036134 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.036143 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.036167 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:36 crc kubenswrapper[4861]: E0315 00:07:36.038650 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:36Z is after 2026-02-23T05:33:13Z" node="crc" Mar 15 00:07:36 crc kubenswrapper[4861]: W0315 00:07:36.118106 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:36Z is after 2026-02-23T05:33:13Z Mar 15 00:07:36 crc kubenswrapper[4861]: E0315 00:07:36.118184 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:36Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 15 00:07:36 crc kubenswrapper[4861]: I0315 00:07:36.333279 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:36Z is after 2026-02-23T05:33:13Z Mar 15 00:07:37 crc kubenswrapper[4861]: I0315 00:07:37.334834 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:37Z is after 2026-02-23T05:33:13Z Mar 15 00:07:37 crc kubenswrapper[4861]: E0315 00:07:37.475277 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:38 crc kubenswrapper[4861]: I0315 00:07:38.333668 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:38Z is after 2026-02-23T05:33:13Z Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.336230 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:07:39Z is after 2026-02-23T05:33:13Z Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.409059 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.410715 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.410798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.410835 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:39 crc kubenswrapper[4861]: I0315 00:07:39.411801 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:39 crc kubenswrapper[4861]: E0315 00:07:39.412100 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.331828 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.332068 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.333177 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.333270 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.333308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:40 crc kubenswrapper[4861]: I0315 00:07:40.338311 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:41 crc kubenswrapper[4861]: I0315 00:07:41.336219 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.634830 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e342fd8c2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,LastTimestamp:2026-03-15 00:06:47.328708802 +0000 UTC m=+0.640906385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.642084 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.648275 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.654499 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.658719 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3ca168b7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.470368951 +0000 UTC m=+0.782566494,LastTimestamp:2026-03-15 00:06:47.470368951 +0000 UTC m=+0.782566494,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.663228 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.509919313 +0000 UTC m=+0.822116866,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.669869 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.509943514 +0000 UTC m=+0.822141057,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.676311 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.509955094 +0000 UTC m=+0.822152637,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.682438 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.510786263 +0000 UTC m=+0.822983806,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.689210 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.510813143 +0000 UTC m=+0.823010686,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.696170 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.510827584 +0000 UTC m=+0.823025137,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.706916 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.512242865 +0000 UTC m=+0.824440418,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.711861 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.512262216 +0000 UTC m=+0.824459759,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.716986 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.512275316 +0000 UTC m=+0.824472859,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.723012 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.512510341 +0000 UTC m=+0.824707924,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.727875 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.512730756 +0000 UTC m=+0.824928299,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.732307 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.512757826 +0000 UTC m=+0.824955369,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.738964 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.513159056 +0000 UTC m=+0.825356609,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.748943 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.513178386 +0000 UTC m=+0.825375939,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.755277 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.513194147 +0000 UTC m=+0.825391690,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.763053 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.51335389 +0000 UTC m=+0.825551413,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.771224 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.513382451 +0000 UTC m=+0.825579984,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.775733 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38057276\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38057276 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393038966 +0000 UTC m=+0.705236539,LastTimestamp:2026-03-15 00:06:47.513391681 +0000 UTC m=+0.825589214,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.783864 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e38047b45\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e38047b45 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.392975685 +0000 UTC m=+0.705173268,LastTimestamp:2026-03-15 00:06:47.513452002 +0000 UTC m=+0.825649545,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.790783 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189cdb0e3805253c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189cdb0e3805253c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.393019196 +0000 UTC m=+0.705216769,LastTimestamp:2026-03-15 00:06:47.513470183 +0000 UTC m=+0.825667736,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.796061 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0e56f0f7ca openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.911790538 +0000 UTC m=+1.223988121,LastTimestamp:2026-03-15 00:06:47.911790538 +0000 UTC m=+1.223988121,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.798011 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0e57378ba8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.916415912 +0000 UTC m=+1.228613485,LastTimestamp:2026-03-15 00:06:47.916415912 +0000 UTC m=+1.228613485,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.802553 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e575eb578 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.91898252 +0000 UTC m=+1.231180093,LastTimestamp:2026-03-15 00:06:47.91898252 +0000 UTC m=+1.231180093,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.806964 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0e57ec394e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.928256846 +0000 UTC m=+1.240454369,LastTimestamp:2026-03-15 00:06:47.928256846 +0000 UTC m=+1.240454369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.813435 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0e585aae9d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:47.935495837 +0000 UTC m=+1.247693380,LastTimestamp:2026-03-15 00:06:47.935495837 +0000 UTC m=+1.247693380,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.818743 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e768fb7b5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.442288053 +0000 UTC m=+1.754485586,LastTimestamp:2026-03-15 00:06:48.442288053 +0000 UTC m=+1.754485586,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.823606 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0e76902462 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.442315874 +0000 UTC m=+1.754513407,LastTimestamp:2026-03-15 00:06:48.442315874 +0000 UTC m=+1.754513407,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.828576 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0e76a13abd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.443435709 +0000 UTC m=+1.755633242,LastTimestamp:2026-03-15 00:06:48.443435709 +0000 UTC m=+1.755633242,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.832477 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0e76ae39f6 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.444287478 +0000 UTC m=+1.756485011,LastTimestamp:2026-03-15 00:06:48.444287478 +0000 UTC m=+1.756485011,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.836965 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0e76b448db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.444684507 +0000 UTC m=+1.756882030,LastTimestamp:2026-03-15 00:06:48.444684507 +0000 UTC m=+1.756882030,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.841394 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e772639b5 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.452151733 +0000 UTC m=+1.764349276,LastTimestamp:2026-03-15 00:06:48.452151733 +0000 UTC m=+1.764349276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.846441 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e77395fd9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.453406681 +0000 UTC m=+1.765604214,LastTimestamp:2026-03-15 00:06:48.453406681 +0000 UTC m=+1.765604214,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.850257 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0e7777dc25 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.457501733 +0000 UTC m=+1.769699266,LastTimestamp:2026-03-15 00:06:48.457501733 +0000 UTC m=+1.769699266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.855446 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0e77899623 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.458663459 +0000 UTC m=+1.770860992,LastTimestamp:2026-03-15 00:06:48.458663459 +0000 UTC m=+1.770860992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.859203 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0e778afdd5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.458755541 +0000 UTC m=+1.770953074,LastTimestamp:2026-03-15 00:06:48.458755541 +0000 UTC m=+1.770953074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.863114 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0e778b9246 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.458793542 +0000 UTC m=+1.770991075,LastTimestamp:2026-03-15 00:06:48.458793542 +0000 UTC m=+1.770991075,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.867297 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e8d097182 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.819364226 +0000 UTC m=+2.131561759,LastTimestamp:2026-03-15 00:06:48.819364226 +0000 UTC m=+2.131561759,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.871623 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e8dbb901a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.831037466 +0000 UTC m=+2.143234999,LastTimestamp:2026-03-15 00:06:48.831037466 +0000 UTC m=+2.143234999,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.876373 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e8dcd7c5e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.832212062 +0000 UTC m=+2.144409605,LastTimestamp:2026-03-15 00:06:48.832212062 +0000 UTC m=+2.144409605,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.882007 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e9993f141 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.029767489 +0000 UTC m=+2.341965022,LastTimestamp:2026-03-15 00:06:49.029767489 +0000 UTC m=+2.341965022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.886882 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e9ac3826f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.049662063 +0000 UTC m=+2.361859586,LastTimestamp:2026-03-15 00:06:49.049662063 +0000 UTC m=+2.361859586,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.891186 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e9ade42a2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.051415202 +0000 UTC m=+2.363612775,LastTimestamp:2026-03-15 00:06:49.051415202 +0000 UTC m=+2.363612775,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.895770 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0ea893f5e2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.281426914 +0000 UTC m=+2.593624497,LastTimestamp:2026-03-15 00:06:49.281426914 +0000 UTC m=+2.593624497,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.899864 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0ea9a14109 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.299075337 +0000 UTC m=+2.611272920,LastTimestamp:2026-03-15 00:06:49.299075337 +0000 UTC m=+2.611272920,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.903748 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0eb14cf3d0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.427768272 +0000 UTC m=+2.739965835,LastTimestamp:2026-03-15 00:06:49.427768272 +0000 UTC m=+2.739965835,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.908456 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0eb1738b9b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.430297499 +0000 UTC m=+2.742495032,LastTimestamp:2026-03-15 00:06:49.430297499 +0000 UTC m=+2.742495032,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.916167 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0eb1bbd3da openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.435034586 +0000 UTC m=+2.747232159,LastTimestamp:2026-03-15 00:06:49.435034586 +0000 UTC m=+2.747232159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.921395 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0eb1c65543 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.435723075 +0000 UTC m=+2.747920648,LastTimestamp:2026-03-15 00:06:49.435723075 +0000 UTC m=+2.747920648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.926866 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ebf5d894d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.663736141 +0000 UTC m=+2.975933674,LastTimestamp:2026-03-15 00:06:49.663736141 +0000 UTC m=+2.975933674,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.932050 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0ebf83464c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.666209356 +0000 UTC m=+2.978406889,LastTimestamp:2026-03-15 00:06:49.666209356 +0000 UTC m=+2.978406889,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.936737 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0ebf8d020b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.666847243 +0000 UTC m=+2.979044776,LastTimestamp:2026-03-15 00:06:49.666847243 +0000 UTC m=+2.979044776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.940512 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ebf9171bd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.667137981 +0000 UTC m=+2.979335514,LastTimestamp:2026-03-15 00:06:49.667137981 +0000 UTC m=+2.979335514,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.944484 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ec03d41f6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.678397942 +0000 UTC m=+2.990595475,LastTimestamp:2026-03-15 00:06:49.678397942 +0000 UTC m=+2.990595475,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.948692 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ec04f27ff openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.679570943 +0000 UTC m=+2.991768476,LastTimestamp:2026-03-15 00:06:49.679570943 +0000 UTC m=+2.991768476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.952231 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189cdb0ec076d3bd openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.682170813 +0000 UTC m=+2.994368356,LastTimestamp:2026-03-15 00:06:49.682170813 +0000 UTC m=+2.994368356,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.955790 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0ec0965116 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.684234518 +0000 UTC m=+2.996432051,LastTimestamp:2026-03-15 00:06:49.684234518 +0000 UTC m=+2.996432051,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.959186 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ec09f7166 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.684832614 +0000 UTC m=+2.997030147,LastTimestamp:2026-03-15 00:06:49.684832614 +0000 UTC m=+2.997030147,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.962897 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ec0c79d13 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.687465235 +0000 UTC m=+2.999662768,LastTimestamp:2026-03-15 00:06:49.687465235 +0000 UTC m=+2.999662768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.967137 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ecbebd4f4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.874388212 +0000 UTC m=+3.186585745,LastTimestamp:2026-03-15 00:06:49.874388212 +0000 UTC m=+3.186585745,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.970869 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ecbff385e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.875658846 +0000 UTC m=+3.187856389,LastTimestamp:2026-03-15 00:06:49.875658846 +0000 UTC m=+3.187856389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.974648 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ecc84c77a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.88441177 +0000 UTC m=+3.196609293,LastTimestamp:2026-03-15 00:06:49.88441177 +0000 UTC m=+3.196609293,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.978025 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ecc952712 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.885484818 +0000 UTC m=+3.197682351,LastTimestamp:2026-03-15 00:06:49.885484818 +0000 UTC m=+3.197682351,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.981802 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0eccbf5fe0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.888251872 +0000 UTC m=+3.200449405,LastTimestamp:2026-03-15 00:06:49.888251872 +0000 UTC m=+3.200449405,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.985513 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0eccd1b0df openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:49.889452255 +0000 UTC m=+3.201649778,LastTimestamp:2026-03-15 00:06:49.889452255 +0000 UTC m=+3.201649778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.989332 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ed8df5814 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.09167362 +0000 UTC m=+3.403871143,LastTimestamp:2026-03-15 00:06:50.09167362 +0000 UTC m=+3.403871143,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.992707 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0ed9227d9a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.096074138 +0000 UTC m=+3.408271671,LastTimestamp:2026-03-15 00:06:50.096074138 +0000 UTC m=+3.408271671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.996378 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ed9d6e511 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.107897105 +0000 UTC m=+3.420094648,LastTimestamp:2026-03-15 00:06:50.107897105 +0000 UTC m=+3.420094648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:41 crc kubenswrapper[4861]: E0315 00:07:41.999817 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ed9e76e4e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.108980814 +0000 UTC m=+3.421178347,LastTimestamp:2026-03-15 00:06:50.108980814 +0000 UTC m=+3.421178347,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.003377 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189cdb0eda1dead7 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.112551639 +0000 UTC m=+3.424749172,LastTimestamp:2026-03-15 00:06:50.112551639 +0000 UTC m=+3.424749172,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.005394 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ee70892f4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.329256692 +0000 UTC m=+3.641454255,LastTimestamp:2026-03-15 00:06:50.329256692 +0000 UTC m=+3.641454255,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.006709 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ee7e633f0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.34378136 +0000 UTC m=+3.655978913,LastTimestamp:2026-03-15 00:06:50.34378136 +0000 UTC m=+3.655978913,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.009412 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ee804385e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.345748574 +0000 UTC m=+3.657946137,LastTimestamp:2026-03-15 00:06:50.345748574 +0000 UTC m=+3.657946137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.013751 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0eeea02352 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.456630098 +0000 UTC m=+3.768827631,LastTimestamp:2026-03-15 00:06:50.456630098 +0000 UTC m=+3.768827631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.017611 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ef47e6587 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.555082119 +0000 UTC m=+3.867279652,LastTimestamp:2026-03-15 00:06:50.555082119 +0000 UTC m=+3.867279652,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.021094 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ef5220842 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.565806146 +0000 UTC m=+3.878003679,LastTimestamp:2026-03-15 00:06:50.565806146 +0000 UTC m=+3.878003679,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.024505 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0efd5f903f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.704056383 +0000 UTC m=+4.016253916,LastTimestamp:2026-03-15 00:06:50.704056383 +0000 UTC m=+4.016253916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.028188 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0efe639544 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.721097028 +0000 UTC m=+4.033294611,LastTimestamp:2026-03-15 00:06:50.721097028 +0000 UTC m=+4.033294611,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.031720 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f2ae534da openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.46778953 +0000 UTC m=+4.779987103,LastTimestamp:2026-03-15 00:06:51.46778953 +0000 UTC m=+4.779987103,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.035452 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f399344c7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.714077895 +0000 UTC m=+5.026275458,LastTimestamp:2026-03-15 00:06:51.714077895 +0000 UTC m=+5.026275458,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.039450 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f3a5eb410 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.727410192 +0000 UTC m=+5.039607725,LastTimestamp:2026-03-15 00:06:51.727410192 +0000 UTC m=+5.039607725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.042533 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f3a748004 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.72883866 +0000 UTC m=+5.041036223,LastTimestamp:2026-03-15 00:06:51.72883866 +0000 UTC m=+5.041036223,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.046245 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f497c1081 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.980992641 +0000 UTC m=+5.293190204,LastTimestamp:2026-03-15 00:06:51.980992641 +0000 UTC m=+5.293190204,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.049585 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f4a6b27f6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.99666175 +0000 UTC m=+5.308859313,LastTimestamp:2026-03-15 00:06:51.99666175 +0000 UTC m=+5.308859313,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.052551 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f4a84bf62 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:51.998338914 +0000 UTC m=+5.310536457,LastTimestamp:2026-03-15 00:06:51.998338914 +0000 UTC m=+5.310536457,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.056844 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f59e1c4f1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.256093425 +0000 UTC m=+5.568290998,LastTimestamp:2026-03-15 00:06:52.256093425 +0000 UTC m=+5.568290998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.060342 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f5af0ec78 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.2738638 +0000 UTC m=+5.586061363,LastTimestamp:2026-03-15 00:06:52.2738638 +0000 UTC m=+5.586061363,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.063705 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f5b0efb8f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.275833743 +0000 UTC m=+5.588031306,LastTimestamp:2026-03-15 00:06:52.275833743 +0000 UTC m=+5.588031306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.067501 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f6b9e9229 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.553679401 +0000 UTC m=+5.865876964,LastTimestamp:2026-03-15 00:06:52.553679401 +0000 UTC m=+5.865876964,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.071506 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f6c9c6c50 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.570315856 +0000 UTC m=+5.882513419,LastTimestamp:2026-03-15 00:06:52.570315856 +0000 UTC m=+5.882513419,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.074639 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f6cbb3ecc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.57233582 +0000 UTC m=+5.884533393,LastTimestamp:2026-03-15 00:06:52.57233582 +0000 UTC m=+5.884533393,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.078791 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f7b9eca7d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.822129277 +0000 UTC m=+6.134326850,LastTimestamp:2026-03-15 00:06:52.822129277 +0000 UTC m=+6.134326850,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.082312 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189cdb0f7c988129 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:52.838494505 +0000 UTC m=+6.150692048,LastTimestamp:2026-03-15 00:06:52.838494505 +0000 UTC m=+6.150692048,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.086001 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-controller-manager-crc.189cdb0fb70e54d9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 15 00:07:42 crc kubenswrapper[4861]: body: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:53.819294937 +0000 UTC m=+7.131492500,LastTimestamp:2026-03-15 00:06:53.819294937 +0000 UTC m=+7.131492500,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.089444 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0fb70f29fb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:53.819349499 +0000 UTC m=+7.131547072,LastTimestamp:2026-03-15 00:06:53.819349499 +0000 UTC m=+7.131547072,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.098736 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c05cc3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 15 00:07:42 crc kubenswrapper[4861]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:42 crc kubenswrapper[4861]: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.598813379 +0000 UTC m=+14.911010922,LastTimestamp:2026-03-15 00:07:01.598813379 +0000 UTC m=+14.911010922,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.102181 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c126cc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.5988651 +0000 UTC m=+14.911062643,LastTimestamp:2026-03-15 00:07:01.5988651 +0000 UTC m=+14.911062643,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.107742 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189cdb1186c05cc3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c05cc3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 15 00:07:42 crc kubenswrapper[4861]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:42 crc kubenswrapper[4861]: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.598813379 +0000 UTC m=+14.911010922,LastTimestamp:2026-03-15 00:07:01.606860534 +0000 UTC m=+14.919058077,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.113207 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189cdb1186c126cc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c126cc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.5988651 +0000 UTC m=+14.911062643,LastTimestamp:2026-03-15 00:07:01.606898315 +0000 UTC m=+14.919095858,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.117077 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189cdb1186c05cc3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c05cc3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 15 00:07:42 crc kubenswrapper[4861]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 15 00:07:42 crc kubenswrapper[4861]: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.598813379 +0000 UTC m=+14.911010922,LastTimestamp:2026-03-15 00:07:01.662637835 +0000 UTC m=+14.974835368,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.123471 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189cdb1186c126cc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb1186c126cc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:01.5988651 +0000 UTC m=+14.911062643,LastTimestamp:2026-03-15 00:07:01.662679176 +0000 UTC m=+14.974876709,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.127217 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189cdb0ee804385e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189cdb0ee804385e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:50.345748574 +0000 UTC m=+3.657946137,LastTimestamp:2026-03-15 00:07:02.52873341 +0000 UTC m=+15.840930983,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.133783 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1b82a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 15 00:07:42 crc kubenswrapper[4861]: body: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819379364 +0000 UTC m=+17.131576897,LastTimestamp:2026-03-15 00:07:03.819379364 +0000 UTC m=+17.131576897,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.139935 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1c6248 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819436616 +0000 UTC m=+17.131634149,LastTimestamp:2026-03-15 00:07:03.819436616 +0000 UTC m=+17.131634149,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.145342 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb120b1b82a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1b82a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 15 00:07:42 crc kubenswrapper[4861]: body: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819379364 +0000 UTC m=+17.131576897,LastTimestamp:2026-03-15 00:07:13.819614968 +0000 UTC m=+27.131812561,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.149308 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb120b1c6248\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1c6248 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819436616 +0000 UTC m=+17.131634149,LastTimestamp:2026-03-15 00:07:13.819730322 +0000 UTC m=+27.131927895,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.153427 4861 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb145f73539d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:13.824355229 +0000 UTC m=+27.136552832,LastTimestamp:2026-03-15 00:07:13.824355229 +0000 UTC m=+27.136552832,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.157389 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb0e77395fd9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e77395fd9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.453406681 +0000 UTC m=+1.765604214,LastTimestamp:2026-03-15 00:07:13.950700206 +0000 UTC m=+27.262897749,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.161238 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb0e8d097182\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e8d097182 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.819364226 +0000 UTC m=+2.131561759,LastTimestamp:2026-03-15 00:07:14.211638795 +0000 UTC m=+27.523836328,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.167604 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb0e8dbb901a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb0e8dbb901a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:06:48.831037466 +0000 UTC m=+2.143234999,LastTimestamp:2026-03-15 00:07:14.223934136 +0000 UTC m=+27.536131669,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.174777 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb120b1b82a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1b82a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 15 00:07:42 crc kubenswrapper[4861]: body: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819379364 +0000 UTC m=+17.131576897,LastTimestamp:2026-03-15 00:07:23.819796984 +0000 UTC m=+37.131994537,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.179353 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb120b1c6248\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1c6248 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819436616 +0000 UTC m=+17.131634149,LastTimestamp:2026-03-15 00:07:23.819864556 +0000 UTC m=+37.132062109,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.186172 4861 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189cdb120b1b82a4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 15 00:07:42 crc kubenswrapper[4861]: &Event{ObjectMeta:{kube-controller-manager-crc.189cdb120b1b82a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 15 00:07:42 crc kubenswrapper[4861]: body: Mar 15 00:07:42 crc kubenswrapper[4861]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:07:03.819379364 +0000 UTC m=+17.131576897,LastTimestamp:2026-03-15 00:07:33.819944662 +0000 UTC m=+47.132142235,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 15 00:07:42 crc kubenswrapper[4861]: > Mar 15 00:07:42 crc kubenswrapper[4861]: I0315 00:07:42.336924 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:42 crc kubenswrapper[4861]: W0315 00:07:42.678214 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:42 crc kubenswrapper[4861]: E0315 00:07:42.678285 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 15 00:07:43 crc kubenswrapper[4861]: E0315 00:07:43.029633 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.039249 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.041158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.041214 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.041232 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.041265 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:43 crc kubenswrapper[4861]: E0315 00:07:43.048348 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.338970 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.819298 4861 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.819385 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.819459 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.819690 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.821522 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.821598 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:43 crc kubenswrapper[4861]: I0315 00:07:43.821612 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:44 crc kubenswrapper[4861]: I0315 00:07:44.335693 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:46 crc kubenswrapper[4861]: I0315 00:07:46.683023 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 15 00:07:46 crc kubenswrapper[4861]: I0315 00:07:46.683199 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb" gracePeriod=30 Mar 15 00:07:46 crc kubenswrapper[4861]: I0315 00:07:46.684591 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.336213 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:47 crc kubenswrapper[4861]: E0315 00:07:47.475863 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.679190 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680261 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680730 4861 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb" exitCode=255 Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680793 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb"} Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680848 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c"} Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680870 4861 scope.go:117] "RemoveContainer" containerID="05096063dec17ba94862090c3de166475493bab9b627ee5046fbfa4928ecdb28" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.680951 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.681621 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.681654 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:47 crc kubenswrapper[4861]: I0315 00:07:47.681669 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.336682 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.687706 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.689941 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.691187 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.691236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:48 crc kubenswrapper[4861]: I0315 00:07:48.691250 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:49 crc kubenswrapper[4861]: I0315 00:07:49.340990 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:50 crc kubenswrapper[4861]: E0315 00:07:50.036060 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.049079 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.050195 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.050227 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.050238 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.050258 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:50 crc kubenswrapper[4861]: E0315 00:07:50.055064 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.335829 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.819423 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.819719 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.824428 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.824801 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.824806 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:50 crc kubenswrapper[4861]: I0315 00:07:50.824845 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.334210 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.697275 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.697743 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.698175 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.698212 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:51 crc kubenswrapper[4861]: I0315 00:07:51.698223 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.334691 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.408847 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.409977 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.410039 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.410062 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.410841 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.702974 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.704970 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940"} Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.705050 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.706117 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.706173 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:52 crc kubenswrapper[4861]: I0315 00:07:52.706185 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.334839 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.708154 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.708739 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.710216 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" exitCode=255 Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.710323 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940"} Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.710413 4861 scope.go:117] "RemoveContainer" containerID="d8e3ac7493e4862d8ad09b0c4cbdaa1abfa9a0447a66335370dea70ad90f66ad" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.710638 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.711424 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.711452 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.711461 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:53 crc kubenswrapper[4861]: I0315 00:07:53.711954 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:07:53 crc kubenswrapper[4861]: E0315 00:07:53.712097 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.278674 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.334525 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.714100 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.721782 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.722667 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.722703 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.722714 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:54 crc kubenswrapper[4861]: I0315 00:07:54.723325 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:07:54 crc kubenswrapper[4861]: E0315 00:07:54.723499 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.334347 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.723959 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.725331 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.725385 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.725398 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:55 crc kubenswrapper[4861]: I0315 00:07:55.725945 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:07:55 crc kubenswrapper[4861]: E0315 00:07:55.726095 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:56 crc kubenswrapper[4861]: I0315 00:07:56.336583 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:56 crc kubenswrapper[4861]: I0315 00:07:56.408463 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:56 crc kubenswrapper[4861]: I0315 00:07:56.409767 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:56 crc kubenswrapper[4861]: I0315 00:07:56.409881 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:56 crc kubenswrapper[4861]: I0315 00:07:56.409942 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:57 crc kubenswrapper[4861]: E0315 00:07:57.044834 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.055981 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.057135 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.057184 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.057198 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.057225 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:07:57 crc kubenswrapper[4861]: E0315 00:07:57.062409 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.335486 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:57 crc kubenswrapper[4861]: E0315 00:07:57.476858 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.516452 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.516719 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.517818 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.517869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.517881 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.518635 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:07:57 crc kubenswrapper[4861]: E0315 00:07:57.518836 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.728652 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 15 00:07:57 crc kubenswrapper[4861]: I0315 00:07:57.744876 4861 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 15 00:07:58 crc kubenswrapper[4861]: I0315 00:07:58.334013 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:07:59 crc kubenswrapper[4861]: I0315 00:07:59.335754 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:00 crc kubenswrapper[4861]: I0315 00:08:00.338427 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:01 crc kubenswrapper[4861]: I0315 00:08:01.335727 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:02 crc kubenswrapper[4861]: I0315 00:08:02.335895 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:03 crc kubenswrapper[4861]: I0315 00:08:03.339797 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:03 crc kubenswrapper[4861]: W0315 00:08:03.794986 4861 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 15 00:08:03 crc kubenswrapper[4861]: E0315 00:08:03.795061 4861 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 15 00:08:04 crc kubenswrapper[4861]: E0315 00:08:04.051353 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.062485 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.063576 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.063607 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.063621 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.063647 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:08:04 crc kubenswrapper[4861]: E0315 00:08:04.067233 4861 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.337243 4861 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.352193 4861 csr.go:261] certificate signing request csr-4rh9t is approved, waiting to be issued Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.362757 4861 csr.go:257] certificate signing request csr-4rh9t is issued Mar 15 00:08:04 crc kubenswrapper[4861]: I0315 00:08:04.384628 4861 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 15 00:08:05 crc kubenswrapper[4861]: I0315 00:08:05.167188 4861 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 15 00:08:05 crc kubenswrapper[4861]: I0315 00:08:05.364173 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-26 04:02:53.90673854 +0000 UTC Mar 15 00:08:05 crc kubenswrapper[4861]: I0315 00:08:05.364239 4861 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6867h54m48.542502784s for next certificate rotation Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:05.783225 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:05.783415 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:05.784639 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:05.784666 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:05.784676 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:06 crc kubenswrapper[4861]: I0315 00:08:06.938654 4861 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 15 00:08:07 crc kubenswrapper[4861]: E0315 00:08:07.477351 4861 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 15 00:08:08 crc kubenswrapper[4861]: I0315 00:08:08.408522 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:08:08 crc kubenswrapper[4861]: I0315 00:08:08.410297 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:08 crc kubenswrapper[4861]: I0315 00:08:08.410358 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:08 crc kubenswrapper[4861]: I0315 00:08:08.410381 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:08 crc kubenswrapper[4861]: I0315 00:08:08.411608 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:08:08 crc kubenswrapper[4861]: E0315 00:08:08.411990 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.029457 4861 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.687370 4861 apiserver.go:52] "Watching apiserver" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.694725 4861 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.695194 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.695978 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.696007 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.696150 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.696235 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.696536 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.696847 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.697150 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.697188 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.697248 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702146 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702223 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702245 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702255 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702352 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702407 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702679 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702804 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.702149 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.736625 4861 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.740292 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745778 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745842 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745877 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745905 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745941 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.745972 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746007 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746045 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746080 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746112 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746142 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746292 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746174 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746882 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746917 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746948 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.746982 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747011 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747037 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747068 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747096 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747128 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747159 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747193 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747238 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747284 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747317 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747330 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747354 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747419 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747418 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747453 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747491 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747522 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747526 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747582 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747621 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747659 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747670 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747687 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747749 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747783 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747807 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747831 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747826 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747871 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747903 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747950 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747961 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748065 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748120 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748026 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748759 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748772 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748805 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748820 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748808 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.748843 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749094 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749115 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749163 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749242 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749374 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749418 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.747856 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749506 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749600 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749639 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749674 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749709 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749742 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749778 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749812 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749847 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749879 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749911 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749945 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749977 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750010 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750040 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750073 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750106 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750137 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750229 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750271 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750305 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750340 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750371 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750454 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750489 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750521 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750577 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750612 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750646 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750677 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750709 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750747 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750785 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750818 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750850 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750883 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750915 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750946 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750979 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751015 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751048 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751080 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751115 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751147 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751180 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751213 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751245 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751279 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751317 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751352 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751387 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751420 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751455 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751494 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751528 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751588 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751627 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751661 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751694 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751728 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751761 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751795 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751830 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751863 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751896 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751934 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751969 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752002 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752037 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752070 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752105 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752227 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752303 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752341 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752378 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752413 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752448 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752482 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752516 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753066 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753400 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753442 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753478 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753707 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753757 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753795 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753830 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753863 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753898 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753932 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753967 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754005 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754121 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754160 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754196 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754282 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754323 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754357 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754393 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754429 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754463 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754498 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754532 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754695 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754737 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754772 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754807 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754845 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754882 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754917 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754950 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754988 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755023 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755056 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755094 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755129 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755164 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755199 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755238 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755274 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755330 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755375 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755412 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755448 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755481 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755518 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755580 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755613 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755647 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755891 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755933 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755969 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756005 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756162 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756205 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756264 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756304 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756341 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756380 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756418 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756453 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756489 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756528 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756586 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756627 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749437 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749688 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749616 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749732 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749946 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.749956 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750033 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767285 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767314 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750042 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750066 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.750726 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751038 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751150 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767858 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751151 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751286 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751414 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751595 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.751597 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767742 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752083 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752106 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752140 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752147 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752251 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752601 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752777 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.752791 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753356 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753581 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753708 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753768 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.753427 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754172 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754264 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754328 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754374 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754390 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754416 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754655 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754704 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754750 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754771 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.754943 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755242 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755198 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755278 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755732 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755773 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755219 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.755819 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756110 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756309 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756326 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756604 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756694 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756726 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.756761 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.757221 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.757242 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.757262 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.757969 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:10.256633163 +0000 UTC m=+83.568830806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.759286 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.759457 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.759859 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.760006 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.760517 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.760594 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.760935 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761067 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761106 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761141 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761470 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761694 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761907 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.761993 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762152 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762184 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762264 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762267 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762384 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762544 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762485 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762609 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.762948 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763112 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763146 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763124 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763207 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763258 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763432 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763433 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763628 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763653 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763679 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763686 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.763783 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.764493 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.764822 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.764841 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.764867 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.765358 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.765391 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.765787 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.765807 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.766174 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.766731 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.766748 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.766798 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767783 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.767888 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.768081 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.770438 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.770385 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771013 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771227 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771303 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771303 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771368 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771460 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771469 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771498 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771721 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771777 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771815 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771865 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771907 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771938 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.771987 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772003 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772263 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772447 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772464 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772623 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772625 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772634 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772646 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772745 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772932 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.772986 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.773664 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.773743 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.773943 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774052 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774128 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774138 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774214 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774228 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774366 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774535 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774716 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774773 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774938 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.774967 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:10.27494753 +0000 UTC m=+83.587145063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774972 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774981 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774994 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775035 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.774872 4861 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775129 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775185 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775066 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775205 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775384 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775490 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775499 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775579 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775598 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775638 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775728 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775764 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775791 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775822 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775918 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.775929 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.776035 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.776282 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.776596 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.776761 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.776822 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.777074 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.777132 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:10.277119996 +0000 UTC m=+83.589317669 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777303 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777490 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777593 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777621 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777640 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777659 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777678 4861 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777696 4861 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777714 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777732 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777752 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777771 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777789 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777807 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777825 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777842 4861 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777860 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777879 4861 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777897 4861 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777916 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777934 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777951 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777969 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.777986 4861 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778003 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778022 4861 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778039 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778056 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778074 4861 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778092 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778109 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778126 4861 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778143 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778162 4861 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778179 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778198 4861 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778214 4861 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778231 4861 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778250 4861 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778273 4861 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778296 4861 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778319 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778343 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778362 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778380 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778397 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778414 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778431 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778449 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778466 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778485 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778501 4861 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778518 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778535 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778552 4861 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778658 4861 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778675 4861 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778693 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778711 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778730 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778748 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778764 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778780 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778797 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778813 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778830 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778848 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778865 4861 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778884 4861 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778902 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778923 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778942 4861 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778958 4861 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778975 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.778992 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779009 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779027 4861 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779044 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779063 4861 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779080 4861 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779097 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779116 4861 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779132 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779149 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779165 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779183 4861 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779201 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779218 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779236 4861 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779254 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779271 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779267 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779289 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779308 4861 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779327 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779345 4861 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779362 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779379 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779395 4861 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779413 4861 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779430 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779446 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779467 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779485 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779501 4861 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779519 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779536 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779554 4861 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779598 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779617 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779636 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779655 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779673 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779691 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779706 4861 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779725 4861 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779743 4861 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779746 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779760 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779838 4861 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779862 4861 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779884 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779902 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779919 4861 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779938 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779955 4861 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779973 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.779993 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780011 4861 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780029 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780045 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780062 4861 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780080 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780128 4861 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780147 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780165 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780185 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780205 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780223 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780241 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780258 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780275 4861 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780293 4861 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780316 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780341 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780365 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780388 4861 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780410 4861 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780433 4861 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780457 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780482 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780507 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780532 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780590 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780616 4861 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780640 4861 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780658 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780675 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780694 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780712 4861 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780729 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780745 4861 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780763 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780780 4861 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780796 4861 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780816 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780843 4861 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780866 4861 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780888 4861 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.780927 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.781472 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.781657 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.797535 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.797596 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.797620 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.797696 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:10.29767059 +0000 UTC m=+83.609868133 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.798048 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.798069 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.798082 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:09 crc kubenswrapper[4861]: E0315 00:08:09.798124 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:10.298110904 +0000 UTC m=+83.610308447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.798201 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.798436 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.798828 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.802854 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.803220 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.803354 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.803175 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.803415 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.804333 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.804636 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.805488 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.806080 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.808685 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.810858 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.811260 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.811354 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.811415 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.817073 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.819943 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.822751 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.830494 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.830655 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.838094 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.845209 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.853183 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881654 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881806 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881855 4861 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881870 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881882 4861 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881895 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881906 4861 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881918 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881930 4861 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881943 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881955 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881967 4861 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881979 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.881991 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882004 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882015 4861 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882028 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882040 4861 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882078 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882102 4861 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882114 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882125 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882135 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882145 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882154 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882164 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882172 4861 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882183 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882192 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882197 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882202 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882239 4861 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 15 00:08:09 crc kubenswrapper[4861]: I0315 00:08:09.882236 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.021675 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.033687 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.034844 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: source /etc/kubernetes/apiserver-url.env Mar 15 00:08:10 crc kubenswrapper[4861]: else Mar 15 00:08:10 crc kubenswrapper[4861]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 15 00:08:10 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.036080 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.043166 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 15 00:08:10 crc kubenswrapper[4861]: W0315 00:08:10.054223 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-fa655f8aa883a1d02d5272585476605adfa5f6f0262cc7c1437af14daa21f497 WatchSource:0}: Error finding container fa655f8aa883a1d02d5272585476605adfa5f6f0262cc7c1437af14daa21f497: Status 404 returned error can't find the container with id fa655f8aa883a1d02d5272585476605adfa5f6f0262cc7c1437af14daa21f497 Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.058195 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: W0315 00:08:10.059043 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-e8b9ee1dbd379019670d1a3f7a3bc41d2b8580874d22bbee604db2c0654ba383 WatchSource:0}: Error finding container e8b9ee1dbd379019670d1a3f7a3bc41d2b8580874d22bbee604db2c0654ba383: Status 404 returned error can't find the container with id e8b9ee1dbd379019670d1a3f7a3bc41d2b8580874d22bbee604db2c0654ba383 Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.059394 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.061989 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:10 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 15 00:08:10 crc kubenswrapper[4861]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 15 00:08:10 crc kubenswrapper[4861]: ho_enable="--enable-hybrid-overlay" Mar 15 00:08:10 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 15 00:08:10 crc kubenswrapper[4861]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 15 00:08:10 crc kubenswrapper[4861]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-host=127.0.0.1 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-port=9743 \ Mar 15 00:08:10 crc kubenswrapper[4861]: ${ho_enable} \ Mar 15 00:08:10 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:10 crc kubenswrapper[4861]: --disable-approver \ Mar 15 00:08:10 crc kubenswrapper[4861]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --wait-for-kubernetes-api=200s \ Mar 15 00:08:10 crc kubenswrapper[4861]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.065059 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:10 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: Mar 15 00:08:10 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --disable-webhook \ Mar 15 00:08:10 crc kubenswrapper[4861]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.066227 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.286549 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.286638 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.286676 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.286786 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.286833 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:11.286817863 +0000 UTC m=+84.599015396 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.287156 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:11.287148613 +0000 UTC m=+84.599346146 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.287187 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.287208 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:11.287203154 +0000 UTC m=+84.599400687 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.387620 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.387747 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387828 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387871 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387899 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387929 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387958 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387983 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.387996 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:11.387966479 +0000 UTC m=+84.700164052 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.388054 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:11.388031271 +0000 UTC m=+84.700228844 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.776107 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e8b9ee1dbd379019670d1a3f7a3bc41d2b8580874d22bbee604db2c0654ba383"} Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.777824 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fa655f8aa883a1d02d5272585476605adfa5f6f0262cc7c1437af14daa21f497"} Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.778450 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:10 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 15 00:08:10 crc kubenswrapper[4861]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 15 00:08:10 crc kubenswrapper[4861]: ho_enable="--enable-hybrid-overlay" Mar 15 00:08:10 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 15 00:08:10 crc kubenswrapper[4861]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 15 00:08:10 crc kubenswrapper[4861]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-host=127.0.0.1 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --webhook-port=9743 \ Mar 15 00:08:10 crc kubenswrapper[4861]: ${ho_enable} \ Mar 15 00:08:10 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:10 crc kubenswrapper[4861]: --disable-approver \ Mar 15 00:08:10 crc kubenswrapper[4861]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --wait-for-kubernetes-api=200s \ Mar 15 00:08:10 crc kubenswrapper[4861]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.779512 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.780408 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:10 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: Mar 15 00:08:10 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:10 crc kubenswrapper[4861]: --disable-webhook \ Mar 15 00:08:10 crc kubenswrapper[4861]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 15 00:08:10 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.780611 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.781347 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"54461b7a4c8f44465afd2dd2a8a1fd9d9829d9c0a00ffef4d8491c2fae4b1d1c"} Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.781605 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.782908 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:10 crc kubenswrapper[4861]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:10 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:10 crc kubenswrapper[4861]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 15 00:08:10 crc kubenswrapper[4861]: source /etc/kubernetes/apiserver-url.env Mar 15 00:08:10 crc kubenswrapper[4861]: else Mar 15 00:08:10 crc kubenswrapper[4861]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 15 00:08:10 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:10 crc kubenswrapper[4861]: fi Mar 15 00:08:10 crc kubenswrapper[4861]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 15 00:08:10 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:10 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:10 crc kubenswrapper[4861]: E0315 00:08:10.784171 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.791716 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.805530 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.820457 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.831140 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.843812 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.853542 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.867797 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.883090 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.895874 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.907866 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.922536 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:10 crc kubenswrapper[4861]: I0315 00:08:10.933615 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.068126 4861 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.069906 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.069964 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.069986 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.070086 4861 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.080999 4861 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.081125 4861 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.082499 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.082587 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.082617 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.082644 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.082667 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.105376 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.110210 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.110287 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.110308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.110333 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.110365 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.126686 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.131021 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.131063 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.131073 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.131088 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.131104 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.146261 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.150635 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.150696 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.150718 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.150743 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.150761 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.165796 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.169721 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.169756 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.169768 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.169798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.169812 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.184890 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.185062 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.186588 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.186640 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.186653 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.186665 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.186674 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.288845 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.288879 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.288892 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.288908 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.288919 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.295230 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.295286 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.295325 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.295400 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.295443 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:13.295429355 +0000 UTC m=+86.607626898 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.295911 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.295927 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:13.295915529 +0000 UTC m=+86.608113072 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.296064 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:13.296039513 +0000 UTC m=+86.608237086 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.392398 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.392459 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.392481 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.392510 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.392533 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.395789 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.395896 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.395987 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396030 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396055 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396055 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396120 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396142 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396154 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:13.396124627 +0000 UTC m=+86.708322200 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.396207 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:13.396184259 +0000 UTC m=+86.708381832 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.408183 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.408225 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.408189 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.408489 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.408523 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:11 crc kubenswrapper[4861]: E0315 00:08:11.408549 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.415257 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.416440 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.419660 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.421191 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.423508 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.424803 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.426326 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.428404 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.429719 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.431651 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.432751 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.434991 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.436079 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.437228 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.439227 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.440399 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.442127 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.443646 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.444288 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.445313 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.445779 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.446712 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.447134 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.448182 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.448611 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.449249 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.450246 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.450717 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.451663 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.452133 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.453130 4861 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.453350 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.456936 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.458700 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.459753 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.463193 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.464547 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.466540 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.468069 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.470357 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.471399 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.473485 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.474902 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.477050 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.478177 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.480241 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.481802 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.484220 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.485409 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.487378 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.488438 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.490483 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.491829 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.492919 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.495241 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.495296 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.495312 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.495334 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.495352 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.598584 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.598632 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.598643 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.598663 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.598681 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.701357 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.701420 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.701438 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.701463 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.701482 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.804430 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.804521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.804595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.804647 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.804666 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.907772 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.907852 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.907870 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.907895 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:11 crc kubenswrapper[4861]: I0315 00:08:11.907912 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:11Z","lastTransitionTime":"2026-03-15T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.011080 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.011166 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.011183 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.011206 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.011225 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.114375 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.114417 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.114428 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.114445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.114456 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.216537 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.216633 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.216652 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.216679 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.216698 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.318981 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.319047 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.319063 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.319088 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.319108 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.421748 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.421814 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.421833 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.421856 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.421876 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.523350 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.523377 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.523387 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.523399 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.523407 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.625689 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.625749 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.625766 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.625787 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.625803 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.727609 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.727685 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.727703 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.727726 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.727743 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.830195 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.830236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.830245 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.830258 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.830267 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.932272 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.932309 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.932318 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.932330 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:12 crc kubenswrapper[4861]: I0315 00:08:12.932338 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:12Z","lastTransitionTime":"2026-03-15T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.035058 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.035097 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.035122 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.035143 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.035157 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.137544 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.137741 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.137768 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.137843 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.137925 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.240510 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.240582 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.240595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.240612 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.240623 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.313800 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.313938 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.314038 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:17.31399146 +0000 UTC m=+90.626189053 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.314113 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.314219 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:17.314184795 +0000 UTC m=+90.626382338 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.314266 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.314331 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:17.314314459 +0000 UTC m=+90.626512042 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.314115 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.343322 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.343356 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.343367 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.343383 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.343397 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.408693 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.408809 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.408947 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.408962 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.409050 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.409241 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.415445 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.415790 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.415819 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.415831 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.415887 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:17.415868277 +0000 UTC m=+90.728065820 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.416264 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.416433 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.416607 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.416727 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:13 crc kubenswrapper[4861]: E0315 00:08:13.416831 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:17.416797285 +0000 UTC m=+90.728994858 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.447406 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.447491 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.447519 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.447544 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.447599 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.551324 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.551806 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.552358 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.552833 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.553367 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.656959 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.657341 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.657623 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.657843 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.658038 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.761445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.761516 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.761539 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.761596 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.761614 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.864882 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.864944 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.864966 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.864995 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.865018 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.968243 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.968349 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.968370 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.968442 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:13 crc kubenswrapper[4861]: I0315 00:08:13.968462 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:13Z","lastTransitionTime":"2026-03-15T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.071992 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.072031 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.072043 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.072058 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.072069 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.174822 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.174870 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.174882 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.174901 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.174914 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.277661 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.277721 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.277743 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.277772 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.277792 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.381604 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.381694 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.381717 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.381746 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.381767 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.485638 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.485697 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.485716 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.485740 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.485757 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.589236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.589584 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.589604 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.589626 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.589644 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.692884 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.692920 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.692932 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.692948 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.692958 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.795100 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.795139 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.795151 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.795167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.795177 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.897689 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.897729 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.897741 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.897757 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:14 crc kubenswrapper[4861]: I0315 00:08:14.897767 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:14Z","lastTransitionTime":"2026-03-15T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.000638 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.000675 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.000683 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.000698 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.000710 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.103614 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.103653 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.103661 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.103675 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.103685 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.206650 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.206696 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.206705 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.206720 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.206729 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.309225 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.309276 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.309289 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.309306 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.309322 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.408811 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:15 crc kubenswrapper[4861]: E0315 00:08:15.408967 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.409006 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.409132 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:15 crc kubenswrapper[4861]: E0315 00:08:15.409196 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:15 crc kubenswrapper[4861]: E0315 00:08:15.409371 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.411862 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.412077 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.412088 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.412102 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.412115 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.514526 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.514595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.514606 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.514621 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.514631 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.616528 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.616578 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.616590 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.616603 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.616611 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.718432 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.718509 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.718518 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.718533 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.718542 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.820245 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.820289 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.820300 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.820317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.820331 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.922659 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.922724 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.922734 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.922758 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:15 crc kubenswrapper[4861]: I0315 00:08:15.922777 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:15Z","lastTransitionTime":"2026-03-15T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.025082 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.025118 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.025126 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.025139 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.025147 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.127106 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.127149 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.127158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.127173 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.127182 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.229429 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.229514 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.229533 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.229591 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.229609 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.331880 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.331944 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.331962 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.331987 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.332005 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.433742 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.433807 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.433829 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.433856 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.433877 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.536127 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.536402 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.536508 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.536663 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.536784 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.649299 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.649467 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.649726 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.649896 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.650059 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.753143 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.753203 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.753220 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.753244 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.753260 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.855882 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.855913 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.855920 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.855934 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.855942 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.958798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.958825 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.958833 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.958845 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:16 crc kubenswrapper[4861]: I0315 00:08:16.958853 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:16Z","lastTransitionTime":"2026-03-15T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.062399 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.062430 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.062439 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.062451 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.062459 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.164693 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.164917 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.165014 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.165075 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.165130 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.267764 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.268008 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.268066 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.268122 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.268177 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.353381 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.353804 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.353992 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.354253 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.354441 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:25.354417686 +0000 UTC m=+98.666615239 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.355058 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.355287 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:25.355267772 +0000 UTC m=+98.667465315 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.355427 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:25.355400586 +0000 UTC m=+98.667598109 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.370127 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.370413 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.370525 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.370662 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.370778 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.408394 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.408507 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.408728 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.408896 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.408902 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.408967 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.422254 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.432209 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.443002 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.454304 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.455054 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.455341 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.455283 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.455695 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.455831 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.455989 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:25.455972475 +0000 UTC m=+98.768170018 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.455513 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.457169 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.457280 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:17 crc kubenswrapper[4861]: E0315 00:08:17.457467 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:25.457452079 +0000 UTC m=+98.769649632 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.462817 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.473417 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.473463 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.473474 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.473486 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.473496 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.474848 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.575796 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.575823 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.575836 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.575850 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.575858 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.678493 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.678529 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.678540 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.678572 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.678583 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.780863 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.781129 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.781282 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.781381 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.781471 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.883574 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.883846 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.883927 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.884016 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.884094 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.986044 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.986603 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.986631 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.986649 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:17 crc kubenswrapper[4861]: I0315 00:08:17.986660 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:17Z","lastTransitionTime":"2026-03-15T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.089151 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.089198 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.089208 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.089226 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.089238 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.191772 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.191842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.191875 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.191902 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.191924 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.294485 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.294526 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.294539 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.294589 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.294607 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.396747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.396788 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.396799 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.396814 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.396823 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.500086 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.500125 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.500141 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.500161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.500173 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.602775 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.602823 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.602840 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.602863 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.602879 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.705868 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.705907 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.705919 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.705935 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.705947 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.808610 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.808729 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.808747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.808770 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.808793 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.911337 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.911412 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.911432 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.911457 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:18 crc kubenswrapper[4861]: I0315 00:08:18.911474 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:18Z","lastTransitionTime":"2026-03-15T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.015050 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.015125 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.015149 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.015177 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.015201 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.117747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.117825 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.117843 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.117869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.117891 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.220645 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.220704 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.220722 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.220747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.220762 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.324510 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.324616 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.324634 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.324658 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.324683 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.408535 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:19 crc kubenswrapper[4861]: E0315 00:08:19.408704 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.408864 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:19 crc kubenswrapper[4861]: E0315 00:08:19.409175 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.409197 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:19 crc kubenswrapper[4861]: E0315 00:08:19.409820 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.427638 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.427680 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.427698 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.427721 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.427738 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.530199 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.530486 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.530666 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.530799 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.530884 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.634138 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.634396 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.634515 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.634632 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.634730 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.737368 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.737429 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.737449 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.737475 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.737493 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.839205 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.839234 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.839244 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.839258 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.839268 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.942475 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.942771 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.942881 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.942959 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:19 crc kubenswrapper[4861]: I0315 00:08:19.943034 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:19Z","lastTransitionTime":"2026-03-15T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.045839 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.046508 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.046621 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.046709 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.047296 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.150856 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.150898 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.150910 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.150925 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.150936 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.253844 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.253897 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.253907 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.253921 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.253931 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.357016 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.357074 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.357083 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.357097 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.357106 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.422906 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:08:20 crc kubenswrapper[4861]: E0315 00:08:20.423107 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.423785 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.459233 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.459484 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.459582 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.459694 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.459803 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.562603 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.562879 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.563159 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.563369 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.563589 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.665980 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.666236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.666325 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.666449 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.666594 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.768989 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.769025 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.769035 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.769050 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.769060 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.803524 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:08:20 crc kubenswrapper[4861]: E0315 00:08:20.803863 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.871095 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.871130 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.871141 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.871155 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.871165 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.973194 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.973380 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.973446 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.973536 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:20 crc kubenswrapper[4861]: I0315 00:08:20.973630 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:20Z","lastTransitionTime":"2026-03-15T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.076068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.076101 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.076110 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.076122 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.076131 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.178323 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.178393 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.178413 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.178443 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.178464 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.281811 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.281944 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.281969 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.282005 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.282026 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.385045 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.385119 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.385136 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.385158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.385179 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.408977 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.409158 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.409348 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.409392 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.409732 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.409927 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.412034 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:21 crc kubenswrapper[4861]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:21 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:21 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:21 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:21 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:21 crc kubenswrapper[4861]: fi Mar 15 00:08:21 crc kubenswrapper[4861]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 15 00:08:21 crc kubenswrapper[4861]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 15 00:08:21 crc kubenswrapper[4861]: ho_enable="--enable-hybrid-overlay" Mar 15 00:08:21 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 15 00:08:21 crc kubenswrapper[4861]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 15 00:08:21 crc kubenswrapper[4861]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 15 00:08:21 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:21 crc kubenswrapper[4861]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 15 00:08:21 crc kubenswrapper[4861]: --webhook-host=127.0.0.1 \ Mar 15 00:08:21 crc kubenswrapper[4861]: --webhook-port=9743 \ Mar 15 00:08:21 crc kubenswrapper[4861]: ${ho_enable} \ Mar 15 00:08:21 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:21 crc kubenswrapper[4861]: --disable-approver \ Mar 15 00:08:21 crc kubenswrapper[4861]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 15 00:08:21 crc kubenswrapper[4861]: --wait-for-kubernetes-api=200s \ Mar 15 00:08:21 crc kubenswrapper[4861]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 15 00:08:21 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:21 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:21 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.415325 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:21 crc kubenswrapper[4861]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:21 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:21 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:21 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:21 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:21 crc kubenswrapper[4861]: fi Mar 15 00:08:21 crc kubenswrapper[4861]: Mar 15 00:08:21 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 15 00:08:21 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:21 crc kubenswrapper[4861]: --disable-webhook \ Mar 15 00:08:21 crc kubenswrapper[4861]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 15 00:08:21 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:21 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:21 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.416418 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.479718 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.479771 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.479785 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.479805 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.479816 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.491995 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.497146 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.497189 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.497204 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.497226 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.497240 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.508507 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.512510 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.512549 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.512613 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.512633 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.512645 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.523437 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.527731 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.527788 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.527799 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.527822 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.527838 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.538462 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.541940 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.541989 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.542001 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.542023 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.542037 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.553707 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:21 crc kubenswrapper[4861]: E0315 00:08:21.553853 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.556044 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.556086 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.556095 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.556111 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.556124 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.658475 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.658538 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.658577 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.658609 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.658627 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.761246 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.761304 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.761312 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.761334 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.761345 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.864103 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.864140 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.864150 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.864166 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.864179 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.967051 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.967134 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.967164 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.967194 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:21 crc kubenswrapper[4861]: I0315 00:08:21.967217 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:21Z","lastTransitionTime":"2026-03-15T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.071442 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.071498 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.071514 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.071537 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.071555 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.174993 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.175055 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.175074 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.175101 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.175120 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.278446 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.278504 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.278521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.278546 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.278607 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.383753 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.383789 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.383798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.383812 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.383824 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: E0315 00:08:22.410649 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:22 crc kubenswrapper[4861]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:22 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:22 crc kubenswrapper[4861]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 15 00:08:22 crc kubenswrapper[4861]: source /etc/kubernetes/apiserver-url.env Mar 15 00:08:22 crc kubenswrapper[4861]: else Mar 15 00:08:22 crc kubenswrapper[4861]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 15 00:08:22 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:22 crc kubenswrapper[4861]: fi Mar 15 00:08:22 crc kubenswrapper[4861]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 15 00:08:22 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:22 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:22 crc kubenswrapper[4861]: E0315 00:08:22.412071 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.486250 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.486516 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.486645 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.486747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.486839 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.589869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.589942 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.589966 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.589999 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.590023 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.693830 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.693891 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.693906 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.693933 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.693947 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.797181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.797234 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.797244 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.797258 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.797303 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.900279 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.900331 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.900349 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.900368 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:22 crc kubenswrapper[4861]: I0315 00:08:22.900380 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:22Z","lastTransitionTime":"2026-03-15T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.003595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.003646 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.003657 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.003679 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.003694 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.105925 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.106009 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.106029 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.106062 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.106085 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.209434 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.209493 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.209503 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.209521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.209532 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.312235 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.312288 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.312297 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.312311 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.312320 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.409063 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.409189 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:23 crc kubenswrapper[4861]: E0315 00:08:23.409338 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:23 crc kubenswrapper[4861]: E0315 00:08:23.410090 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.410821 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:23 crc kubenswrapper[4861]: E0315 00:08:23.411400 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.414973 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.415023 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.415079 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.415109 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.415129 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.517674 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.517764 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.517779 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.517798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.517810 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.620278 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.620322 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.620333 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.620347 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.620356 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.723173 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.723245 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.723264 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.723302 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.723341 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.825526 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.825666 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.825692 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.825717 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.825733 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.928069 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.928116 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.928131 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.928149 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:23 crc kubenswrapper[4861]: I0315 00:08:23.928161 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:23Z","lastTransitionTime":"2026-03-15T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.030421 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.030468 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.030480 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.030496 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.030505 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.134253 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.134338 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.134367 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.134402 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.134422 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.237623 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.237672 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.237682 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.237699 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.237709 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.341286 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.341340 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.341351 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.341370 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.341384 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: E0315 00:08:24.410717 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:24 crc kubenswrapper[4861]: E0315 00:08:24.411980 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.444591 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.444630 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.444655 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.444673 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.444682 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.547735 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.547811 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.547825 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.547853 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.547870 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.650594 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.650638 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.650651 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.650677 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.650693 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.753305 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.753337 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.753347 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.753540 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.753552 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.855690 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.855776 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.855788 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.855803 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.855812 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.958451 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.958494 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.958502 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.958524 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:24 crc kubenswrapper[4861]: I0315 00:08:24.958534 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:24Z","lastTransitionTime":"2026-03-15T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.061030 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.061142 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.061153 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.061165 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.061174 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.162881 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.162917 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.162928 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.162943 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.162952 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.266164 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.266204 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.266213 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.266229 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.266239 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.368824 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.368879 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.368888 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.368902 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.368911 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.408166 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.408204 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.408296 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.408346 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.408413 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.408489 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.428982 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.429069 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.429111 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.429233 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:08:41.429200949 +0000 UTC m=+114.741398492 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.429259 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.429313 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:41.429296652 +0000 UTC m=+114.741494275 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.429322 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.429369 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:41.429360165 +0000 UTC m=+114.741557708 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.471139 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.471167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.471175 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.471187 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.471198 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.529752 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.529806 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.529915 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.529929 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.529941 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.529988 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:41.529975434 +0000 UTC m=+114.842172967 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.530529 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.530694 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.530794 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:25 crc kubenswrapper[4861]: E0315 00:08:25.530933 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:41.530913422 +0000 UTC m=+114.843110975 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.573459 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.573494 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.573503 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.573515 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.573524 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.675280 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.675567 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.675722 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.675800 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.675868 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.777926 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.777973 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.777987 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.778007 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.778019 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.881342 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.881804 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.882008 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.882153 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.882296 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.985015 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.985291 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.985382 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.985478 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:25 crc kubenswrapper[4861]: I0315 00:08:25.985579 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:25Z","lastTransitionTime":"2026-03-15T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.087594 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.087644 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.087660 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.087680 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.087696 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.190851 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.190902 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.190915 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.190933 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.190947 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.322126 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.322185 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.322207 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.322236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.322257 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.425655 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.425690 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.425699 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.425712 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.425720 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.527705 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.527736 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.527744 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.527756 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.527795 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.630533 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.630612 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.630628 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.630649 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.630664 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.733647 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.733705 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.733723 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.733745 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.733759 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.835871 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.835913 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.835925 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.835940 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.835952 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.938377 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.938423 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.938435 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.938452 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:26 crc kubenswrapper[4861]: I0315 00:08:26.938463 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:26Z","lastTransitionTime":"2026-03-15T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.040770 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.040806 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.040815 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.040827 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.040835 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.142943 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.143001 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.143017 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.143037 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.143052 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.244830 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.244873 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.244882 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.244905 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.244915 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.347614 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.347652 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.347663 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.347676 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.347686 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.408741 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:27 crc kubenswrapper[4861]: E0315 00:08:27.408890 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.408931 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.408976 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:27 crc kubenswrapper[4861]: E0315 00:08:27.409078 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:27 crc kubenswrapper[4861]: E0315 00:08:27.409168 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.419026 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.426717 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.436010 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.445543 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.449189 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.449234 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.449247 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.449264 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.449275 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.453248 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.460962 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.467428 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.552452 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.552510 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.552526 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.552548 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.552587 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.606488 4861 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.656175 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.656222 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.656233 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.656250 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.656266 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.758168 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.758218 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.758235 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.758252 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.758264 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.860261 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.860299 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.860308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.860322 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.860332 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.962818 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.962868 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.962885 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.962904 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:27 crc kubenswrapper[4861]: I0315 00:08:27.962921 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:27Z","lastTransitionTime":"2026-03-15T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.065899 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.065945 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.065956 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.065974 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.065986 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.168269 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.168308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.168319 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.168336 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.168346 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.269890 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.269926 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.269933 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.269945 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.269953 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.371703 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.371747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.371755 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.371770 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.371781 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.474263 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.474306 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.474317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.474336 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.474349 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.576651 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.576686 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.576695 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.576707 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.576715 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.683579 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.683629 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.683642 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.683661 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.683674 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.785585 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.785625 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.785636 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.785650 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.785661 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.887161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.887191 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.887200 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.887212 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.887220 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.989973 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.990014 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.990022 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.990036 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:28 crc kubenswrapper[4861]: I0315 00:08:28.990046 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:28Z","lastTransitionTime":"2026-03-15T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.093124 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.093167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.093179 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.093196 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.093216 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.196223 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.196274 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.196285 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.196301 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.196311 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.299123 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.299169 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.299181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.299199 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.299210 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.401429 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.401497 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.401517 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.401540 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.401581 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.408610 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:29 crc kubenswrapper[4861]: E0315 00:08:29.408723 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.408611 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:29 crc kubenswrapper[4861]: E0315 00:08:29.408800 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.408609 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:29 crc kubenswrapper[4861]: E0315 00:08:29.408853 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.504346 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.504395 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.504407 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.504425 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.504439 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.607395 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.607473 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.607487 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.607505 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.607520 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.709928 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.709967 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.709979 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.709996 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.710007 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.812998 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.813033 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.813046 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.813061 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.813072 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.918689 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.918778 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.918804 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.918837 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:29 crc kubenswrapper[4861]: I0315 00:08:29.918862 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:29Z","lastTransitionTime":"2026-03-15T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.022111 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.022174 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.022193 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.022217 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.022265 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.124448 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.124503 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.124517 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.124536 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.124550 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.226842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.226886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.226901 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.226921 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.226936 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.330105 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.330156 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.330171 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.330192 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.330205 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.433338 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.433400 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.433415 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.433435 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.433450 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.536152 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.536201 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.536211 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.536225 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.536235 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.638298 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.638332 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.638342 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.638357 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.638368 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.740771 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.740813 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.740825 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.740847 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.740864 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.843260 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.843290 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.843301 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.843317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.843327 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.945615 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.945677 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.945686 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.945701 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:30 crc kubenswrapper[4861]: I0315 00:08:30.945712 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:30Z","lastTransitionTime":"2026-03-15T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.047930 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.047972 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.047985 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.048000 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.048012 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.150375 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.150421 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.150430 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.150445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.150453 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.252932 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.252961 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.252970 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.252983 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.252992 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.355257 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.355323 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.355349 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.355379 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.355399 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.407929 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.407929 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.408138 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.408449 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.408663 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.408924 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.457669 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.457720 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.457731 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.457747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.457762 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.557195 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.557236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.557246 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.557264 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.557277 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.576222 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.580958 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.581262 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.581641 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.581901 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.582132 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.595613 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-ffqzt"] Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.596441 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.596968 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.605479 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.606182 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.606624 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.608980 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.609055 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.609077 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.609107 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.609131 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.616028 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.625956 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.629746 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.629800 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.629837 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.629858 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.629871 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.630734 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.640537 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.642803 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.644038 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.644187 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.644845 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.645043 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.645186 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.654286 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.657009 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.657148 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.659284 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.659328 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.659345 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.659370 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.659390 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.665307 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.673325 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.685140 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nfg5\" (UniqueName: \"kubernetes.io/projected/3a70c127-00db-4104-a64d-26a45ec0b728-kube-api-access-4nfg5\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.685232 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a70c127-00db-4104-a64d-26a45ec0b728-hosts-file\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.687239 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.701470 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.762310 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.762391 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.762412 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.762444 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.762471 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.785841 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a70c127-00db-4104-a64d-26a45ec0b728-hosts-file\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.786129 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3a70c127-00db-4104-a64d-26a45ec0b728-hosts-file\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.786357 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nfg5\" (UniqueName: \"kubernetes.io/projected/3a70c127-00db-4104-a64d-26a45ec0b728-kube-api-access-4nfg5\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.808236 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nfg5\" (UniqueName: \"kubernetes.io/projected/3a70c127-00db-4104-a64d-26a45ec0b728-kube-api-access-4nfg5\") pod \"node-resolver-ffqzt\" (UID: \"3a70c127-00db-4104-a64d-26a45ec0b728\") " pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.865135 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.865177 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.865191 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.865210 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.865225 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.922615 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ffqzt" Mar 15 00:08:31 crc kubenswrapper[4861]: W0315 00:08:31.935323 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a70c127_00db_4104_a64d_26a45ec0b728.slice/crio-b5e36fe7e4f00d617402c9ce6700092c91f6bbfec8114d2aaa23065289c9ee89 WatchSource:0}: Error finding container b5e36fe7e4f00d617402c9ce6700092c91f6bbfec8114d2aaa23065289c9ee89: Status 404 returned error can't find the container with id b5e36fe7e4f00d617402c9ce6700092c91f6bbfec8114d2aaa23065289c9ee89 Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.939904 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:31 crc kubenswrapper[4861]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:31 crc kubenswrapper[4861]: set -uo pipefail Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 15 00:08:31 crc kubenswrapper[4861]: HOSTS_FILE="/etc/hosts" Mar 15 00:08:31 crc kubenswrapper[4861]: TEMP_FILE="/etc/hosts.tmp" Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: # Make a temporary file with the old hosts file's attributes. Mar 15 00:08:31 crc kubenswrapper[4861]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 15 00:08:31 crc kubenswrapper[4861]: echo "Failed to preserve hosts file. Exiting." Mar 15 00:08:31 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:31 crc kubenswrapper[4861]: fi Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: while true; do Mar 15 00:08:31 crc kubenswrapper[4861]: declare -A svc_ips Mar 15 00:08:31 crc kubenswrapper[4861]: for svc in "${services[@]}"; do Mar 15 00:08:31 crc kubenswrapper[4861]: # Fetch service IP from cluster dns if present. We make several tries Mar 15 00:08:31 crc kubenswrapper[4861]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 15 00:08:31 crc kubenswrapper[4861]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 15 00:08:31 crc kubenswrapper[4861]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 15 00:08:31 crc kubenswrapper[4861]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:31 crc kubenswrapper[4861]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:31 crc kubenswrapper[4861]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:31 crc kubenswrapper[4861]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 15 00:08:31 crc kubenswrapper[4861]: for i in ${!cmds[*]} Mar 15 00:08:31 crc kubenswrapper[4861]: do Mar 15 00:08:31 crc kubenswrapper[4861]: ips=($(eval "${cmds[i]}")) Mar 15 00:08:31 crc kubenswrapper[4861]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 15 00:08:31 crc kubenswrapper[4861]: svc_ips["${svc}"]="${ips[@]}" Mar 15 00:08:31 crc kubenswrapper[4861]: break Mar 15 00:08:31 crc kubenswrapper[4861]: fi Mar 15 00:08:31 crc kubenswrapper[4861]: done Mar 15 00:08:31 crc kubenswrapper[4861]: done Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: # Update /etc/hosts only if we get valid service IPs Mar 15 00:08:31 crc kubenswrapper[4861]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 15 00:08:31 crc kubenswrapper[4861]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 15 00:08:31 crc kubenswrapper[4861]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 15 00:08:31 crc kubenswrapper[4861]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 15 00:08:31 crc kubenswrapper[4861]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 15 00:08:31 crc kubenswrapper[4861]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 15 00:08:31 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:31 crc kubenswrapper[4861]: continue Mar 15 00:08:31 crc kubenswrapper[4861]: fi Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: # Append resolver entries for services Mar 15 00:08:31 crc kubenswrapper[4861]: rc=0 Mar 15 00:08:31 crc kubenswrapper[4861]: for svc in "${!svc_ips[@]}"; do Mar 15 00:08:31 crc kubenswrapper[4861]: for ip in ${svc_ips[${svc}]}; do Mar 15 00:08:31 crc kubenswrapper[4861]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 15 00:08:31 crc kubenswrapper[4861]: done Mar 15 00:08:31 crc kubenswrapper[4861]: done Mar 15 00:08:31 crc kubenswrapper[4861]: if [[ $rc -ne 0 ]]; then Mar 15 00:08:31 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:31 crc kubenswrapper[4861]: continue Mar 15 00:08:31 crc kubenswrapper[4861]: fi Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: Mar 15 00:08:31 crc kubenswrapper[4861]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 15 00:08:31 crc kubenswrapper[4861]: # Replace /etc/hosts with our modified version if needed Mar 15 00:08:31 crc kubenswrapper[4861]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 15 00:08:31 crc kubenswrapper[4861]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 15 00:08:31 crc kubenswrapper[4861]: fi Mar 15 00:08:31 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:31 crc kubenswrapper[4861]: unset svc_ips Mar 15 00:08:31 crc kubenswrapper[4861]: done Mar 15 00:08:31 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nfg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-ffqzt_openshift-dns(3a70c127-00db-4104-a64d-26a45ec0b728): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:31 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:31 crc kubenswrapper[4861]: E0315 00:08:31.944752 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-ffqzt" podUID="3a70c127-00db-4104-a64d-26a45ec0b728" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.949270 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-bbzq6"] Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.949549 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-r8m45"] Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.949767 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bbzq6" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.949798 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rtlks"] Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.949883 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.950991 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.952190 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.954281 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.955903 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.956242 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.956489 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.956752 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.958186 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.958398 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.958457 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.966944 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967040 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967103 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967216 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:31Z","lastTransitionTime":"2026-03-15T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967746 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.967926 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.968207 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.971988 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.981960 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.990088 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:31 crc kubenswrapper[4861]: I0315 00:08:31.997827 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.014780 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.035804 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.052288 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.064265 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.072092 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.072154 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.072173 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.072196 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.072212 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.079699 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090299 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-kubelet\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090439 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-multus-daemon-config\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090485 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-bin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090548 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-hostroot\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090591 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-conf-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090652 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-os-release\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090761 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-system-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090793 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5442c197-f16b-462d-b7ef-f509c8b1e858-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090825 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-multus\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090844 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-etc-kubernetes\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090861 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090878 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-socket-dir-parent\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090893 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-netns\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090906 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-multus-certs\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090920 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86x2p\" (UniqueName: \"kubernetes.io/projected/8cdc187d-d699-431a-9355-5a2268bf3821-kube-api-access-86x2p\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090944 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-cnibin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090959 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5442c197-f16b-462d-b7ef-f509c8b1e858-proxy-tls\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090975 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkwmg\" (UniqueName: \"kubernetes.io/projected/5442c197-f16b-462d-b7ef-f509c8b1e858-kube-api-access-tkwmg\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.090993 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-system-cni-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091009 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-binary-copy\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091027 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5442c197-f16b-462d-b7ef-f509c8b1e858-rootfs\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091070 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-os-release\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091181 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091227 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk5r7\" (UniqueName: \"kubernetes.io/projected/a59701a7-ce44-4f3b-b03b-e01058001ac9-kube-api-access-hk5r7\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091263 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-cnibin\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091289 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091332 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-cni-binary-copy\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.091356 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-k8s-cni-cncf-io\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.107785 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.122894 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.139615 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.155064 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.167930 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.174745 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.174772 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.174781 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.174810 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.174820 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.181989 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192588 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-cnibin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192619 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-multus-certs\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192638 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86x2p\" (UniqueName: \"kubernetes.io/projected/8cdc187d-d699-431a-9355-5a2268bf3821-kube-api-access-86x2p\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192657 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5442c197-f16b-462d-b7ef-f509c8b1e858-proxy-tls\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192674 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkwmg\" (UniqueName: \"kubernetes.io/projected/5442c197-f16b-462d-b7ef-f509c8b1e858-kube-api-access-tkwmg\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192690 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-system-cni-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192715 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5442c197-f16b-462d-b7ef-f509c8b1e858-rootfs\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192734 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-binary-copy\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192740 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-multus-certs\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192824 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-os-release\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192857 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5442c197-f16b-462d-b7ef-f509c8b1e858-rootfs\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192845 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-system-cni-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192707 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-cnibin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.192752 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-os-release\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193046 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193081 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-cnibin\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193134 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193162 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk5r7\" (UniqueName: \"kubernetes.io/projected/a59701a7-ce44-4f3b-b03b-e01058001ac9-kube-api-access-hk5r7\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193186 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-k8s-cni-cncf-io\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193210 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-cni-binary-copy\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193149 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193265 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-kubelet\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193234 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-kubelet\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193327 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193330 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-multus-daemon-config\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193381 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-bin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193405 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-os-release\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193436 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-system-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193457 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-hostroot\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193478 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-conf-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193513 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5442c197-f16b-462d-b7ef-f509c8b1e858-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193539 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193582 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-socket-dir-parent\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193608 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-netns\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193629 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-multus\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193653 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-etc-kubernetes\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193704 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-etc-kubernetes\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193735 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-cnibin\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193795 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-binary-copy\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.193849 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-hostroot\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194018 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-multus-daemon-config\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194099 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-k8s-cni-cncf-io\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194235 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-os-release\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194274 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-bin\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194334 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a59701a7-ce44-4f3b-b03b-e01058001ac9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194399 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-socket-dir-parent\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194436 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-run-netns\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194472 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-host-var-lib-cni-multus\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194506 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-multus-conf-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194667 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8cdc187d-d699-431a-9355-5a2268bf3821-cni-binary-copy\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194723 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8cdc187d-d699-431a-9355-5a2268bf3821-system-cni-dir\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.194762 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a59701a7-ce44-4f3b-b03b-e01058001ac9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.195156 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5442c197-f16b-462d-b7ef-f509c8b1e858-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.199970 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5442c197-f16b-462d-b7ef-f509c8b1e858-proxy-tls\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.209361 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.211626 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk5r7\" (UniqueName: \"kubernetes.io/projected/a59701a7-ce44-4f3b-b03b-e01058001ac9-kube-api-access-hk5r7\") pod \"multus-additional-cni-plugins-rtlks\" (UID: \"a59701a7-ce44-4f3b-b03b-e01058001ac9\") " pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.212102 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86x2p\" (UniqueName: \"kubernetes.io/projected/8cdc187d-d699-431a-9355-5a2268bf3821-kube-api-access-86x2p\") pod \"multus-bbzq6\" (UID: \"8cdc187d-d699-431a-9355-5a2268bf3821\") " pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.220937 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkwmg\" (UniqueName: \"kubernetes.io/projected/5442c197-f16b-462d-b7ef-f509c8b1e858-kube-api-access-tkwmg\") pod \"machine-config-daemon-r8m45\" (UID: \"5442c197-f16b-462d-b7ef-f509c8b1e858\") " pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.228211 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.241239 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.256897 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.275323 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bbzq6" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.278430 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.278487 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.278505 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.278528 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.278542 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.280973 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:08:32 crc kubenswrapper[4861]: W0315 00:08:32.285703 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cdc187d_d699_431a_9355_5a2268bf3821.slice/crio-d7ee353cf1adec6c14922eaa4a71d9f18f61a7ccaa6fb2fc3b5a820c315240d1 WatchSource:0}: Error finding container d7ee353cf1adec6c14922eaa4a71d9f18f61a7ccaa6fb2fc3b5a820c315240d1: Status 404 returned error can't find the container with id d7ee353cf1adec6c14922eaa4a71d9f18f61a7ccaa6fb2fc3b5a820c315240d1 Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.288504 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:32 crc kubenswrapper[4861]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 15 00:08:32 crc kubenswrapper[4861]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 15 00:08:32 crc kubenswrapper[4861]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86x2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:32 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.290574 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.293934 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rtlks" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.295937 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.298933 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.300129 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:08:32 crc kubenswrapper[4861]: W0315 00:08:32.305187 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda59701a7_ce44_4f3b_b03b_e01058001ac9.slice/crio-d6e1c4ebb316a14468d14192093d3176899b9a5337142bcce05aaf2f8cac3778 WatchSource:0}: Error finding container d6e1c4ebb316a14468d14192093d3176899b9a5337142bcce05aaf2f8cac3778: Status 404 returned error can't find the container with id d6e1c4ebb316a14468d14192093d3176899b9a5337142bcce05aaf2f8cac3778 Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.307252 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hk5r7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-rtlks_openshift-multus(a59701a7-ce44-4f3b-b03b-e01058001ac9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.310639 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-rtlks" podUID="a59701a7-ce44-4f3b-b03b-e01058001ac9" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.311516 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pmf25"] Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.312609 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.315270 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.315629 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.316045 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.316298 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.317946 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.319380 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.319416 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.331295 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.343417 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.356235 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.369077 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381261 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381287 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381256 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381297 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381497 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.381512 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394266 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394708 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394741 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394765 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394786 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k452t\" (UniqueName: \"kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394940 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.394993 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395024 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395055 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395104 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395133 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395198 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395248 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395277 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395514 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395585 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395638 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395695 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395801 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395862 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.395915 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.410737 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.424756 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.443631 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.455474 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.469311 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.481658 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.483827 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.483886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.483907 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.483940 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.483962 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497064 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497116 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497151 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497222 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497306 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497393 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497418 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497484 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497499 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497549 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497611 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497629 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497667 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497680 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497715 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497726 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497758 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497788 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497797 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497828 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k452t\" (UniqueName: \"kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497840 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497760 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497880 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497899 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497941 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497976 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.497987 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498021 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498047 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498077 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498113 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498342 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498392 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498620 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498732 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498781 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498787 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.498998 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.504308 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.518036 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k452t\" (UniqueName: \"kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t\") pod \"ovnkube-node-pmf25\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.587633 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.587696 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.587706 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.587728 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.587748 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.627411 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:08:32 crc kubenswrapper[4861]: W0315 00:08:32.641767 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa2d91cf_7d1a_4567_9593_654417784ef1.slice/crio-fed8b2ae383c8884aa99f085a158f86ff2a696c61c4d1e82d0ed69f359da793d WatchSource:0}: Error finding container fed8b2ae383c8884aa99f085a158f86ff2a696c61c4d1e82d0ed69f359da793d: Status 404 returned error can't find the container with id fed8b2ae383c8884aa99f085a158f86ff2a696c61c4d1e82d0ed69f359da793d Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.645148 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:32 crc kubenswrapper[4861]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 15 00:08:32 crc kubenswrapper[4861]: apiVersion: v1 Mar 15 00:08:32 crc kubenswrapper[4861]: clusters: Mar 15 00:08:32 crc kubenswrapper[4861]: - cluster: Mar 15 00:08:32 crc kubenswrapper[4861]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 15 00:08:32 crc kubenswrapper[4861]: server: https://api-int.crc.testing:6443 Mar 15 00:08:32 crc kubenswrapper[4861]: name: default-cluster Mar 15 00:08:32 crc kubenswrapper[4861]: contexts: Mar 15 00:08:32 crc kubenswrapper[4861]: - context: Mar 15 00:08:32 crc kubenswrapper[4861]: cluster: default-cluster Mar 15 00:08:32 crc kubenswrapper[4861]: namespace: default Mar 15 00:08:32 crc kubenswrapper[4861]: user: default-auth Mar 15 00:08:32 crc kubenswrapper[4861]: name: default-context Mar 15 00:08:32 crc kubenswrapper[4861]: current-context: default-context Mar 15 00:08:32 crc kubenswrapper[4861]: kind: Config Mar 15 00:08:32 crc kubenswrapper[4861]: preferences: {} Mar 15 00:08:32 crc kubenswrapper[4861]: users: Mar 15 00:08:32 crc kubenswrapper[4861]: - name: default-auth Mar 15 00:08:32 crc kubenswrapper[4861]: user: Mar 15 00:08:32 crc kubenswrapper[4861]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:32 crc kubenswrapper[4861]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:32 crc kubenswrapper[4861]: EOF Mar 15 00:08:32 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k452t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:32 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.648670 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.699759 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.699826 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.699839 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.699858 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.699869 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.804067 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.804154 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.804171 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.804200 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.804219 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.836256 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"fed8b2ae383c8884aa99f085a158f86ff2a696c61c4d1e82d0ed69f359da793d"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.839692 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"24efcf1516f1bbe1b30acaf2804e78bd28911c89f07ce84ea56d1fea00429e26"} Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.841873 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:32 crc kubenswrapper[4861]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 15 00:08:32 crc kubenswrapper[4861]: apiVersion: v1 Mar 15 00:08:32 crc kubenswrapper[4861]: clusters: Mar 15 00:08:32 crc kubenswrapper[4861]: - cluster: Mar 15 00:08:32 crc kubenswrapper[4861]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 15 00:08:32 crc kubenswrapper[4861]: server: https://api-int.crc.testing:6443 Mar 15 00:08:32 crc kubenswrapper[4861]: name: default-cluster Mar 15 00:08:32 crc kubenswrapper[4861]: contexts: Mar 15 00:08:32 crc kubenswrapper[4861]: - context: Mar 15 00:08:32 crc kubenswrapper[4861]: cluster: default-cluster Mar 15 00:08:32 crc kubenswrapper[4861]: namespace: default Mar 15 00:08:32 crc kubenswrapper[4861]: user: default-auth Mar 15 00:08:32 crc kubenswrapper[4861]: name: default-context Mar 15 00:08:32 crc kubenswrapper[4861]: current-context: default-context Mar 15 00:08:32 crc kubenswrapper[4861]: kind: Config Mar 15 00:08:32 crc kubenswrapper[4861]: preferences: {} Mar 15 00:08:32 crc kubenswrapper[4861]: users: Mar 15 00:08:32 crc kubenswrapper[4861]: - name: default-auth Mar 15 00:08:32 crc kubenswrapper[4861]: user: Mar 15 00:08:32 crc kubenswrapper[4861]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:32 crc kubenswrapper[4861]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:32 crc kubenswrapper[4861]: EOF Mar 15 00:08:32 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k452t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:32 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.843339 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.843537 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.844409 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerStarted","Data":"d6e1c4ebb316a14468d14192093d3176899b9a5337142bcce05aaf2f8cac3778"} Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.846794 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.848348 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.849353 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hk5r7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-rtlks_openshift-multus(a59701a7-ce44-4f3b-b03b-e01058001ac9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.850526 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-rtlks" podUID="a59701a7-ce44-4f3b-b03b-e01058001ac9" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.852695 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerStarted","Data":"d7ee353cf1adec6c14922eaa4a71d9f18f61a7ccaa6fb2fc3b5a820c315240d1"} Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.854471 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:32 crc kubenswrapper[4861]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 15 00:08:32 crc kubenswrapper[4861]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 15 00:08:32 crc kubenswrapper[4861]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86x2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:32 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.854841 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ffqzt" event={"ID":"3a70c127-00db-4104-a64d-26a45ec0b728","Type":"ContainerStarted","Data":"b5e36fe7e4f00d617402c9ce6700092c91f6bbfec8114d2aaa23065289c9ee89"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.855609 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.855806 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.857903 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:32 crc kubenswrapper[4861]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:32 crc kubenswrapper[4861]: set -uo pipefail Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 15 00:08:32 crc kubenswrapper[4861]: HOSTS_FILE="/etc/hosts" Mar 15 00:08:32 crc kubenswrapper[4861]: TEMP_FILE="/etc/hosts.tmp" Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: # Make a temporary file with the old hosts file's attributes. Mar 15 00:08:32 crc kubenswrapper[4861]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 15 00:08:32 crc kubenswrapper[4861]: echo "Failed to preserve hosts file. Exiting." Mar 15 00:08:32 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:32 crc kubenswrapper[4861]: fi Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: while true; do Mar 15 00:08:32 crc kubenswrapper[4861]: declare -A svc_ips Mar 15 00:08:32 crc kubenswrapper[4861]: for svc in "${services[@]}"; do Mar 15 00:08:32 crc kubenswrapper[4861]: # Fetch service IP from cluster dns if present. We make several tries Mar 15 00:08:32 crc kubenswrapper[4861]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 15 00:08:32 crc kubenswrapper[4861]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 15 00:08:32 crc kubenswrapper[4861]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 15 00:08:32 crc kubenswrapper[4861]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:32 crc kubenswrapper[4861]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:32 crc kubenswrapper[4861]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:32 crc kubenswrapper[4861]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 15 00:08:32 crc kubenswrapper[4861]: for i in ${!cmds[*]} Mar 15 00:08:32 crc kubenswrapper[4861]: do Mar 15 00:08:32 crc kubenswrapper[4861]: ips=($(eval "${cmds[i]}")) Mar 15 00:08:32 crc kubenswrapper[4861]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 15 00:08:32 crc kubenswrapper[4861]: svc_ips["${svc}"]="${ips[@]}" Mar 15 00:08:32 crc kubenswrapper[4861]: break Mar 15 00:08:32 crc kubenswrapper[4861]: fi Mar 15 00:08:32 crc kubenswrapper[4861]: done Mar 15 00:08:32 crc kubenswrapper[4861]: done Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: # Update /etc/hosts only if we get valid service IPs Mar 15 00:08:32 crc kubenswrapper[4861]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 15 00:08:32 crc kubenswrapper[4861]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 15 00:08:32 crc kubenswrapper[4861]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 15 00:08:32 crc kubenswrapper[4861]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 15 00:08:32 crc kubenswrapper[4861]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 15 00:08:32 crc kubenswrapper[4861]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 15 00:08:32 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:32 crc kubenswrapper[4861]: continue Mar 15 00:08:32 crc kubenswrapper[4861]: fi Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: # Append resolver entries for services Mar 15 00:08:32 crc kubenswrapper[4861]: rc=0 Mar 15 00:08:32 crc kubenswrapper[4861]: for svc in "${!svc_ips[@]}"; do Mar 15 00:08:32 crc kubenswrapper[4861]: for ip in ${svc_ips[${svc}]}; do Mar 15 00:08:32 crc kubenswrapper[4861]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 15 00:08:32 crc kubenswrapper[4861]: done Mar 15 00:08:32 crc kubenswrapper[4861]: done Mar 15 00:08:32 crc kubenswrapper[4861]: if [[ $rc -ne 0 ]]; then Mar 15 00:08:32 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:32 crc kubenswrapper[4861]: continue Mar 15 00:08:32 crc kubenswrapper[4861]: fi Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: Mar 15 00:08:32 crc kubenswrapper[4861]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 15 00:08:32 crc kubenswrapper[4861]: # Replace /etc/hosts with our modified version if needed Mar 15 00:08:32 crc kubenswrapper[4861]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 15 00:08:32 crc kubenswrapper[4861]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 15 00:08:32 crc kubenswrapper[4861]: fi Mar 15 00:08:32 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:32 crc kubenswrapper[4861]: unset svc_ips Mar 15 00:08:32 crc kubenswrapper[4861]: done Mar 15 00:08:32 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nfg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-ffqzt_openshift-dns(3a70c127-00db-4104-a64d-26a45ec0b728): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:32 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:32 crc kubenswrapper[4861]: E0315 00:08:32.859393 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-ffqzt" podUID="3a70c127-00db-4104-a64d-26a45ec0b728" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.870954 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.887496 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.902158 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.912406 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.912477 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.912499 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.912530 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.912579 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:32Z","lastTransitionTime":"2026-03-15T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.916435 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.933413 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.949334 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.961745 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.972592 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:32 crc kubenswrapper[4861]: I0315 00:08:32.983367 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.004153 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.016653 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.016734 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.016758 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.016793 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.016816 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.021935 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.034626 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.048386 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.058477 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.085366 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.098758 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.112108 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.120190 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.120231 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.120240 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.120257 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.120266 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.124785 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.133205 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.145824 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.161430 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.176517 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.188632 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.223920 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.223967 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.223980 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.224002 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.224014 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.327206 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.327319 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.327338 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.327394 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.327411 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.408336 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.408519 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:33 crc kubenswrapper[4861]: E0315 00:08:33.408724 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.408733 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:33 crc kubenswrapper[4861]: E0315 00:08:33.408835 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:33 crc kubenswrapper[4861]: E0315 00:08:33.410079 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:33 crc kubenswrapper[4861]: E0315 00:08:33.414339 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:33 crc kubenswrapper[4861]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:33 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:33 crc kubenswrapper[4861]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 15 00:08:33 crc kubenswrapper[4861]: source /etc/kubernetes/apiserver-url.env Mar 15 00:08:33 crc kubenswrapper[4861]: else Mar 15 00:08:33 crc kubenswrapper[4861]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 15 00:08:33 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:33 crc kubenswrapper[4861]: fi Mar 15 00:08:33 crc kubenswrapper[4861]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 15 00:08:33 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:33 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:33 crc kubenswrapper[4861]: E0315 00:08:33.415576 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.433241 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.433308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.433326 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.433353 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.433370 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.537267 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.537328 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.537346 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.537371 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.537390 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.641056 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.641130 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.641150 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.641181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.641202 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.744961 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.745044 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.745064 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.745094 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.745116 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.848614 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.848684 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.848701 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.848724 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.848741 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.952344 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.952408 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.952432 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.952458 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:33 crc kubenswrapper[4861]: I0315 00:08:33.952474 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:33Z","lastTransitionTime":"2026-03-15T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.055511 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.055554 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.055596 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.055611 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.055619 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.159760 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.159836 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.159855 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.159886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.159906 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.264129 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.264201 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.264218 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.264248 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.264269 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.367612 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.367686 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.367711 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.367741 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.367761 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.472085 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.472161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.472184 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.472216 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.472242 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.575817 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.575871 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.575886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.575907 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.575921 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.680181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.680257 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.680274 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.680300 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.680321 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.783401 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.783461 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.783474 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.783494 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.783507 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.886675 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.886761 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.886791 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.886832 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.886860 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.989767 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.989842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.989867 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.989905 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:34 crc kubenswrapper[4861]: I0315 00:08:34.989929 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:34Z","lastTransitionTime":"2026-03-15T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.092374 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.092411 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.092421 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.092437 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.092450 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.195832 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.195892 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.195909 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.195934 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.195953 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.298540 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.298635 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.298675 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.298705 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.298729 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.402667 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.402719 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.402737 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.402760 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.402774 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.408168 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.408328 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.408357 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.408497 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.409010 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.409442 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.409304 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.411728 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.412637 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:35 crc kubenswrapper[4861]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:35 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:35 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:35 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:35 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:35 crc kubenswrapper[4861]: fi Mar 15 00:08:35 crc kubenswrapper[4861]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 15 00:08:35 crc kubenswrapper[4861]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 15 00:08:35 crc kubenswrapper[4861]: ho_enable="--enable-hybrid-overlay" Mar 15 00:08:35 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 15 00:08:35 crc kubenswrapper[4861]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 15 00:08:35 crc kubenswrapper[4861]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 15 00:08:35 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:35 crc kubenswrapper[4861]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 15 00:08:35 crc kubenswrapper[4861]: --webhook-host=127.0.0.1 \ Mar 15 00:08:35 crc kubenswrapper[4861]: --webhook-port=9743 \ Mar 15 00:08:35 crc kubenswrapper[4861]: ${ho_enable} \ Mar 15 00:08:35 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:35 crc kubenswrapper[4861]: --disable-approver \ Mar 15 00:08:35 crc kubenswrapper[4861]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 15 00:08:35 crc kubenswrapper[4861]: --wait-for-kubernetes-api=200s \ Mar 15 00:08:35 crc kubenswrapper[4861]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 15 00:08:35 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:35 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:35 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.412879 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.422231 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:35 crc kubenswrapper[4861]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:35 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:35 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:35 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:35 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:35 crc kubenswrapper[4861]: fi Mar 15 00:08:35 crc kubenswrapper[4861]: Mar 15 00:08:35 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 15 00:08:35 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:35 crc kubenswrapper[4861]: --disable-webhook \ Mar 15 00:08:35 crc kubenswrapper[4861]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 15 00:08:35 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:35 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:35 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:35 crc kubenswrapper[4861]: E0315 00:08:35.423429 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.505916 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.505985 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.505998 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.506018 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.506034 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.608166 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.608223 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.608241 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.608266 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.608285 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.711181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.711234 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.711244 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.711265 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.711279 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.814411 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.814477 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.814493 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.814518 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.814530 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.866704 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.868790 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.869856 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.883333 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.897762 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.917503 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.917630 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.917660 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.917696 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.917724 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:35Z","lastTransitionTime":"2026-03-15T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.923006 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.937953 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.951907 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.960224 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.973239 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:35 crc kubenswrapper[4861]: I0315 00:08:35.998410 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.011359 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.019955 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.019985 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.019996 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.020011 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.020024 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.025395 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.044591 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.056692 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.123092 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.123158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.123181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.123210 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.123232 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.225625 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.225677 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.225688 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.225707 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.225718 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.330121 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.330167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.330184 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.330205 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.330218 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.433425 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.433479 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.433491 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.433514 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.433527 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.535884 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.535925 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.535934 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.535948 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.535959 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.638798 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.638838 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.638846 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.638859 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.638868 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.741611 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.741645 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.741654 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.741684 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.741694 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.844032 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.844068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.844077 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.844092 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.844101 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.946589 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.946657 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.946674 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.946699 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:36 crc kubenswrapper[4861]: I0315 00:08:36.946712 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:36Z","lastTransitionTime":"2026-03-15T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.048975 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.049018 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.049026 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.049039 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.049048 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.151418 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.151498 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.151516 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.151544 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.151604 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.255804 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.255928 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.255963 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.255991 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.256011 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.358168 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.358265 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.358291 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.358325 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.358350 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.408764 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.408777 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.408801 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:37 crc kubenswrapper[4861]: E0315 00:08:37.408868 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:37 crc kubenswrapper[4861]: E0315 00:08:37.409012 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:37 crc kubenswrapper[4861]: E0315 00:08:37.409437 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.421852 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.424354 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.450192 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.460892 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.460959 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.460975 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.460997 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.461012 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.464893 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.480315 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.495296 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.508971 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.523728 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.538377 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.550042 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.562741 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.563129 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.563206 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.563231 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.563261 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.563282 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.571714 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.584531 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.665384 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.665445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.665462 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.665485 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.665501 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.768812 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.768874 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.768886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.768904 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.768916 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.870962 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.870992 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.871000 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.871012 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.871020 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.974447 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.974544 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.974591 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.974619 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:37 crc kubenswrapper[4861]: I0315 00:08:37.974636 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:37Z","lastTransitionTime":"2026-03-15T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.078304 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.078388 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.078412 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.078446 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.078686 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.182265 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.182327 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.182344 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.182368 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.182385 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.286521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.286624 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.286651 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.286681 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.286703 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.389742 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.389822 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.389842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.389869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.389889 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.493113 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.493201 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.493219 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.493243 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.493260 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.572909 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tw5n7"] Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.573446 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.575809 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.575810 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.578022 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.578258 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.589042 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.596158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.596210 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.596226 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.596250 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.596269 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.618226 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.633599 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.650097 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.666619 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.671853 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5674aba8-8b60-409d-b473-6ce136bed7d8-serviceca\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.671940 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swv5m\" (UniqueName: \"kubernetes.io/projected/5674aba8-8b60-409d-b473-6ce136bed7d8-kube-api-access-swv5m\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.671973 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5674aba8-8b60-409d-b473-6ce136bed7d8-host\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.676839 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.693775 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.700000 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.700317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.700473 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.700680 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.700897 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.706310 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.717761 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.726990 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.739379 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.770256 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.773188 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5674aba8-8b60-409d-b473-6ce136bed7d8-serviceca\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.773300 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swv5m\" (UniqueName: \"kubernetes.io/projected/5674aba8-8b60-409d-b473-6ce136bed7d8-kube-api-access-swv5m\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.773350 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5674aba8-8b60-409d-b473-6ce136bed7d8-host\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.773514 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5674aba8-8b60-409d-b473-6ce136bed7d8-host\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.775609 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5674aba8-8b60-409d-b473-6ce136bed7d8-serviceca\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.798104 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.803235 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.803271 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.803284 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.803301 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.803313 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.814332 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swv5m\" (UniqueName: \"kubernetes.io/projected/5674aba8-8b60-409d-b473-6ce136bed7d8-kube-api-access-swv5m\") pod \"node-ca-tw5n7\" (UID: \"5674aba8-8b60-409d-b473-6ce136bed7d8\") " pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.823994 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.894721 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tw5n7" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.905792 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.905849 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.905860 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.905881 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:38 crc kubenswrapper[4861]: I0315 00:08:38.905894 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:38Z","lastTransitionTime":"2026-03-15T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:38 crc kubenswrapper[4861]: W0315 00:08:38.906800 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5674aba8_8b60_409d_b473_6ce136bed7d8.slice/crio-9be5a5cece425560da99b5486be75b83e74d3761ff78aa5bdc0878d1518d8d08 WatchSource:0}: Error finding container 9be5a5cece425560da99b5486be75b83e74d3761ff78aa5bdc0878d1518d8d08: Status 404 returned error can't find the container with id 9be5a5cece425560da99b5486be75b83e74d3761ff78aa5bdc0878d1518d8d08 Mar 15 00:08:38 crc kubenswrapper[4861]: E0315 00:08:38.909309 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:38 crc kubenswrapper[4861]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 15 00:08:38 crc kubenswrapper[4861]: while [ true ]; Mar 15 00:08:38 crc kubenswrapper[4861]: do Mar 15 00:08:38 crc kubenswrapper[4861]: for f in $(ls /tmp/serviceca); do Mar 15 00:08:38 crc kubenswrapper[4861]: echo $f Mar 15 00:08:38 crc kubenswrapper[4861]: ca_file_path="/tmp/serviceca/${f}" Mar 15 00:08:38 crc kubenswrapper[4861]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 15 00:08:38 crc kubenswrapper[4861]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 15 00:08:38 crc kubenswrapper[4861]: if [ -e "${reg_dir_path}" ]; then Mar 15 00:08:38 crc kubenswrapper[4861]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 15 00:08:38 crc kubenswrapper[4861]: else Mar 15 00:08:38 crc kubenswrapper[4861]: mkdir $reg_dir_path Mar 15 00:08:38 crc kubenswrapper[4861]: cp $ca_file_path $reg_dir_path/ca.crt Mar 15 00:08:38 crc kubenswrapper[4861]: fi Mar 15 00:08:38 crc kubenswrapper[4861]: done Mar 15 00:08:38 crc kubenswrapper[4861]: for d in $(ls /etc/docker/certs.d); do Mar 15 00:08:38 crc kubenswrapper[4861]: echo $d Mar 15 00:08:38 crc kubenswrapper[4861]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 15 00:08:38 crc kubenswrapper[4861]: reg_conf_path="/tmp/serviceca/${dp}" Mar 15 00:08:38 crc kubenswrapper[4861]: if [ ! -e "${reg_conf_path}" ]; then Mar 15 00:08:38 crc kubenswrapper[4861]: rm -rf /etc/docker/certs.d/$d Mar 15 00:08:38 crc kubenswrapper[4861]: fi Mar 15 00:08:38 crc kubenswrapper[4861]: done Mar 15 00:08:38 crc kubenswrapper[4861]: sleep 60 & wait ${!} Mar 15 00:08:38 crc kubenswrapper[4861]: done Mar 15 00:08:38 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swv5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-tw5n7_openshift-image-registry(5674aba8-8b60-409d-b473-6ce136bed7d8): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:38 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:38 crc kubenswrapper[4861]: E0315 00:08:38.910457 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-tw5n7" podUID="5674aba8-8b60-409d-b473-6ce136bed7d8" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.008532 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.008594 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.008602 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.008617 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.008627 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.110523 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.110832 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.110914 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.111003 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.111100 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.213242 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.213279 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.213289 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.213306 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.213315 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.315920 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.316006 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.316030 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.316062 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.316086 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.408316 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.408356 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.408323 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:39 crc kubenswrapper[4861]: E0315 00:08:39.408480 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:39 crc kubenswrapper[4861]: E0315 00:08:39.408672 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:39 crc kubenswrapper[4861]: E0315 00:08:39.408824 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.418519 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.418626 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.418669 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.418693 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.418708 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.520762 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.520844 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.520857 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.520883 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.520900 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.623521 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.623593 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.623607 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.623627 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.623640 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.726640 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.726720 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.726732 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.726758 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.726773 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.829997 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.830072 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.830095 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.830127 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.830152 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.878088 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tw5n7" event={"ID":"5674aba8-8b60-409d-b473-6ce136bed7d8","Type":"ContainerStarted","Data":"9be5a5cece425560da99b5486be75b83e74d3761ff78aa5bdc0878d1518d8d08"} Mar 15 00:08:39 crc kubenswrapper[4861]: E0315 00:08:39.879636 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:39 crc kubenswrapper[4861]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 15 00:08:39 crc kubenswrapper[4861]: while [ true ]; Mar 15 00:08:39 crc kubenswrapper[4861]: do Mar 15 00:08:39 crc kubenswrapper[4861]: for f in $(ls /tmp/serviceca); do Mar 15 00:08:39 crc kubenswrapper[4861]: echo $f Mar 15 00:08:39 crc kubenswrapper[4861]: ca_file_path="/tmp/serviceca/${f}" Mar 15 00:08:39 crc kubenswrapper[4861]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 15 00:08:39 crc kubenswrapper[4861]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 15 00:08:39 crc kubenswrapper[4861]: if [ -e "${reg_dir_path}" ]; then Mar 15 00:08:39 crc kubenswrapper[4861]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 15 00:08:39 crc kubenswrapper[4861]: else Mar 15 00:08:39 crc kubenswrapper[4861]: mkdir $reg_dir_path Mar 15 00:08:39 crc kubenswrapper[4861]: cp $ca_file_path $reg_dir_path/ca.crt Mar 15 00:08:39 crc kubenswrapper[4861]: fi Mar 15 00:08:39 crc kubenswrapper[4861]: done Mar 15 00:08:39 crc kubenswrapper[4861]: for d in $(ls /etc/docker/certs.d); do Mar 15 00:08:39 crc kubenswrapper[4861]: echo $d Mar 15 00:08:39 crc kubenswrapper[4861]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 15 00:08:39 crc kubenswrapper[4861]: reg_conf_path="/tmp/serviceca/${dp}" Mar 15 00:08:39 crc kubenswrapper[4861]: if [ ! -e "${reg_conf_path}" ]; then Mar 15 00:08:39 crc kubenswrapper[4861]: rm -rf /etc/docker/certs.d/$d Mar 15 00:08:39 crc kubenswrapper[4861]: fi Mar 15 00:08:39 crc kubenswrapper[4861]: done Mar 15 00:08:39 crc kubenswrapper[4861]: sleep 60 & wait ${!} Mar 15 00:08:39 crc kubenswrapper[4861]: done Mar 15 00:08:39 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swv5m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-tw5n7_openshift-image-registry(5674aba8-8b60-409d-b473-6ce136bed7d8): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:39 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:39 crc kubenswrapper[4861]: E0315 00:08:39.880845 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-tw5n7" podUID="5674aba8-8b60-409d-b473-6ce136bed7d8" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.890981 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.899919 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.911108 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.919298 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.929174 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.933282 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.933358 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.933376 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.933407 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.933427 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:39Z","lastTransitionTime":"2026-03-15T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.948151 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.957975 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.968323 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.977902 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.987479 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:39 crc kubenswrapper[4861]: I0315 00:08:39.995704 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.003013 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.018629 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.030624 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.036325 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.036384 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.036399 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.036418 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.036431 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.138547 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.138659 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.138672 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.138691 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.138703 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.240907 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.241034 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.241056 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.241075 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.241086 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.343350 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.343415 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.343437 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.343467 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.343489 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.446042 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.446071 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.446178 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.446194 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.446203 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.548513 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.548573 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.548586 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.548604 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.548618 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.655357 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.655409 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.655418 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.655432 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.655441 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.758293 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.758346 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.758362 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.758385 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.758406 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.862189 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.862272 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.862292 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.862325 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.862345 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.965810 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.965899 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.965954 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.965984 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:40 crc kubenswrapper[4861]: I0315 00:08:40.966002 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:40Z","lastTransitionTime":"2026-03-15T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.069014 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.069109 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.069133 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.069167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.069190 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.172627 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.172726 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.172754 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.172792 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.172813 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.275942 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.276036 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.276063 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.276093 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.276115 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.379780 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.379848 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.379865 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.379891 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.379911 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.408932 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.408932 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.409086 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.409304 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.409451 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.409793 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.483105 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.483164 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.483181 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.483211 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.483235 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.500710 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.500825 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.500890 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:09:13.500863115 +0000 UTC m=+146.813060668 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.500916 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.500953 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.500976 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:13.500958048 +0000 UTC m=+146.813155601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.501113 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.501161 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:13.501147603 +0000 UTC m=+146.813345146 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.586676 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.586767 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.586787 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.586815 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.586835 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.601397 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.601482 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601725 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601753 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601773 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601797 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601842 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601850 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:13.601827925 +0000 UTC m=+146.914025488 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601867 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.601965 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:13.601930548 +0000 UTC m=+146.914128111 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.690174 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.690279 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.690302 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.690336 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.690357 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.794348 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.794422 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.794448 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.794480 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.794504 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.807940 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.807982 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.807991 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.808005 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.808016 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.819403 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.824371 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.824435 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.824453 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.824478 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.824497 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.839534 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.842919 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.842962 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.842977 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.842997 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.843011 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.857657 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.862477 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.862516 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.862528 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.862548 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.862579 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.876833 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.881817 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.881847 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.881857 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.881874 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.881886 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.899252 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:41 crc kubenswrapper[4861]: E0315 00:08:41.899399 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.901649 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.901689 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.901702 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.901724 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:41 crc kubenswrapper[4861]: I0315 00:08:41.901735 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:41Z","lastTransitionTime":"2026-03-15T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.004743 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.004829 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.004851 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.004877 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.004895 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.107308 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.107377 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.107397 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.107422 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.107438 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.215390 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.215460 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.215493 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.215520 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.215542 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.319276 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.319333 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.319350 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.319374 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.319395 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.422093 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.422158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.422179 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.422204 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.422224 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.525934 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.526167 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.526178 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.526197 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.526211 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.629339 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.629770 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.629950 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.630125 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.630309 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.733033 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.733317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.733419 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.733484 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.733548 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.835897 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.835949 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.835958 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.835973 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.835983 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.938062 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.938096 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.938104 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.938116 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:42 crc kubenswrapper[4861]: I0315 00:08:42.938124 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:42Z","lastTransitionTime":"2026-03-15T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.040068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.040107 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.040120 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.040135 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.040144 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.143012 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.143081 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.143104 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.143133 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.143156 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.245634 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.245696 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.245714 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.245737 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.245755 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.349103 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.349161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.349176 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.349198 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.349212 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.408872 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:43 crc kubenswrapper[4861]: E0315 00:08:43.409057 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.409612 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:43 crc kubenswrapper[4861]: E0315 00:08:43.409722 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.409813 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:43 crc kubenswrapper[4861]: E0315 00:08:43.409894 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.452547 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.452610 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.452621 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.452638 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.452651 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.555491 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.555636 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.555657 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.555681 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.555698 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.658484 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.658535 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.658549 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.658595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.658612 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.761205 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.761271 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.761291 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.761315 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.761333 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.864960 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.865040 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.865062 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.865089 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.865112 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.968242 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.968311 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.968325 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.968345 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:43 crc kubenswrapper[4861]: I0315 00:08:43.968358 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:43Z","lastTransitionTime":"2026-03-15T00:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.071633 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.071712 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.071726 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.071747 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.071759 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.174351 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.174652 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.174823 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.174923 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.175013 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.277781 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.277816 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.277824 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.277842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.277850 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.381624 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.381693 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.381712 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.381739 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.381757 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.413270 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl"] Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.413911 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.418324 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.418466 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.425943 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.426002 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.426063 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.426095 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bf8\" (UniqueName: \"kubernetes.io/projected/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-kube-api-access-v7bf8\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.442833 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.453062 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.462988 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.474801 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.484895 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.484992 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.485063 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.485130 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.485185 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.489119 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.500765 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.510854 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.527616 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.527822 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7bf8\" (UniqueName: \"kubernetes.io/projected/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-kube-api-access-v7bf8\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.528001 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.528170 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.529213 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.529393 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.544949 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.545207 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.550398 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7bf8\" (UniqueName: \"kubernetes.io/projected/86b8bbef-bcb6-4395-bd33-1a45a55b25d2-kube-api-access-v7bf8\") pod \"ovnkube-control-plane-749d76644c-r25wl\" (UID: \"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.559053 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.569000 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.581503 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.589465 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.589527 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.589545 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.589631 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.589652 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.592635 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.604258 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.621225 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.639292 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.692847 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.693218 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.693310 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.693411 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.693515 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.735952 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" Mar 15 00:08:44 crc kubenswrapper[4861]: W0315 00:08:44.755426 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86b8bbef_bcb6_4395_bd33_1a45a55b25d2.slice/crio-ba3f147a9e9ac900d8150b33ad600686f3bd8e88b83ca7b6a6ba2768e40a2591 WatchSource:0}: Error finding container ba3f147a9e9ac900d8150b33ad600686f3bd8e88b83ca7b6a6ba2768e40a2591: Status 404 returned error can't find the container with id ba3f147a9e9ac900d8150b33ad600686f3bd8e88b83ca7b6a6ba2768e40a2591 Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.758043 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:44 crc kubenswrapper[4861]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:44 crc kubenswrapper[4861]: set -euo pipefail Mar 15 00:08:44 crc kubenswrapper[4861]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 15 00:08:44 crc kubenswrapper[4861]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 15 00:08:44 crc kubenswrapper[4861]: # As the secret mount is optional we must wait for the files to be present. Mar 15 00:08:44 crc kubenswrapper[4861]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 15 00:08:44 crc kubenswrapper[4861]: TS=$(date +%s) Mar 15 00:08:44 crc kubenswrapper[4861]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 15 00:08:44 crc kubenswrapper[4861]: HAS_LOGGED_INFO=0 Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: log_missing_certs(){ Mar 15 00:08:44 crc kubenswrapper[4861]: CUR_TS=$(date +%s) Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 15 00:08:44 crc kubenswrapper[4861]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 15 00:08:44 crc kubenswrapper[4861]: HAS_LOGGED_INFO=1 Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: } Mar 15 00:08:44 crc kubenswrapper[4861]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 15 00:08:44 crc kubenswrapper[4861]: log_missing_certs Mar 15 00:08:44 crc kubenswrapper[4861]: sleep 5 Mar 15 00:08:44 crc kubenswrapper[4861]: done Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 15 00:08:44 crc kubenswrapper[4861]: exec /usr/bin/kube-rbac-proxy \ Mar 15 00:08:44 crc kubenswrapper[4861]: --logtostderr \ Mar 15 00:08:44 crc kubenswrapper[4861]: --secure-listen-address=:9108 \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 15 00:08:44 crc kubenswrapper[4861]: --upstream=http://127.0.0.1:29108/ \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-private-key-file=${TLS_PK} \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-cert-file=${TLS_CERT} Mar 15 00:08:44 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7bf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-r25wl_openshift-ovn-kubernetes(86b8bbef-bcb6-4395-bd33-1a45a55b25d2): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:44 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.761124 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:44 crc kubenswrapper[4861]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:44 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:44 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_join_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_join_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_transit_switch_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_transit_switch_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: dns_name_resolver_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "false" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: persistent_ips_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "true" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: # This is needed so that converting clusters from GA to TP Mar 15 00:08:44 crc kubenswrapper[4861]: # will rollout control plane pods as well Mar 15 00:08:44 crc kubenswrapper[4861]: network_segmentation_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: multi_network_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "true" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: multi_network_enabled_flag="--enable-multi-network" Mar 15 00:08:44 crc kubenswrapper[4861]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 15 00:08:44 crc kubenswrapper[4861]: exec /usr/bin/ovnkube \ Mar 15 00:08:44 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:44 crc kubenswrapper[4861]: --init-cluster-manager "${K8S_NODE}" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 15 00:08:44 crc kubenswrapper[4861]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-bind-address "127.0.0.1:29108" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-enable-pprof \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-enable-config-duration \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v4_join_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v6_join_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${dns_name_resolver_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${persistent_ips_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${multi_network_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${network_segmentation_enabled_flag} Mar 15 00:08:44 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7bf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-r25wl_openshift-ovn-kubernetes(86b8bbef-bcb6-4395-bd33-1a45a55b25d2): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:44 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.762443 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" podUID="86b8bbef-bcb6-4395-bd33-1a45a55b25d2" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.796866 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.796913 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.796923 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.796940 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.796952 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.891748 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" event={"ID":"86b8bbef-bcb6-4395-bd33-1a45a55b25d2","Type":"ContainerStarted","Data":"ba3f147a9e9ac900d8150b33ad600686f3bd8e88b83ca7b6a6ba2768e40a2591"} Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.894251 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:44 crc kubenswrapper[4861]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:44 crc kubenswrapper[4861]: set -euo pipefail Mar 15 00:08:44 crc kubenswrapper[4861]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 15 00:08:44 crc kubenswrapper[4861]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 15 00:08:44 crc kubenswrapper[4861]: # As the secret mount is optional we must wait for the files to be present. Mar 15 00:08:44 crc kubenswrapper[4861]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 15 00:08:44 crc kubenswrapper[4861]: TS=$(date +%s) Mar 15 00:08:44 crc kubenswrapper[4861]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 15 00:08:44 crc kubenswrapper[4861]: HAS_LOGGED_INFO=0 Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: log_missing_certs(){ Mar 15 00:08:44 crc kubenswrapper[4861]: CUR_TS=$(date +%s) Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 15 00:08:44 crc kubenswrapper[4861]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 15 00:08:44 crc kubenswrapper[4861]: HAS_LOGGED_INFO=1 Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: } Mar 15 00:08:44 crc kubenswrapper[4861]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 15 00:08:44 crc kubenswrapper[4861]: log_missing_certs Mar 15 00:08:44 crc kubenswrapper[4861]: sleep 5 Mar 15 00:08:44 crc kubenswrapper[4861]: done Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 15 00:08:44 crc kubenswrapper[4861]: exec /usr/bin/kube-rbac-proxy \ Mar 15 00:08:44 crc kubenswrapper[4861]: --logtostderr \ Mar 15 00:08:44 crc kubenswrapper[4861]: --secure-listen-address=:9108 \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 15 00:08:44 crc kubenswrapper[4861]: --upstream=http://127.0.0.1:29108/ \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-private-key-file=${TLS_PK} \ Mar 15 00:08:44 crc kubenswrapper[4861]: --tls-cert-file=${TLS_CERT} Mar 15 00:08:44 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7bf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-r25wl_openshift-ovn-kubernetes(86b8bbef-bcb6-4395-bd33-1a45a55b25d2): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:44 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.897903 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:44 crc kubenswrapper[4861]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:44 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:44 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_join_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_join_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_transit_switch_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_transit_switch_subnet_opt= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "" != "" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: dns_name_resolver_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "false" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: persistent_ips_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "true" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: # This is needed so that converting clusters from GA to TP Mar 15 00:08:44 crc kubenswrapper[4861]: # will rollout control plane pods as well Mar 15 00:08:44 crc kubenswrapper[4861]: network_segmentation_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: multi_network_enabled_flag= Mar 15 00:08:44 crc kubenswrapper[4861]: if [[ "true" == "true" ]]; then Mar 15 00:08:44 crc kubenswrapper[4861]: multi_network_enabled_flag="--enable-multi-network" Mar 15 00:08:44 crc kubenswrapper[4861]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 15 00:08:44 crc kubenswrapper[4861]: fi Mar 15 00:08:44 crc kubenswrapper[4861]: Mar 15 00:08:44 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 15 00:08:44 crc kubenswrapper[4861]: exec /usr/bin/ovnkube \ Mar 15 00:08:44 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:44 crc kubenswrapper[4861]: --init-cluster-manager "${K8S_NODE}" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 15 00:08:44 crc kubenswrapper[4861]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-bind-address "127.0.0.1:29108" \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-enable-pprof \ Mar 15 00:08:44 crc kubenswrapper[4861]: --metrics-enable-config-duration \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v4_join_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v6_join_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${dns_name_resolver_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${persistent_ips_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${multi_network_enabled_flag} \ Mar 15 00:08:44 crc kubenswrapper[4861]: ${network_segmentation_enabled_flag} Mar 15 00:08:44 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7bf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-r25wl_openshift-ovn-kubernetes(86b8bbef-bcb6-4395-bd33-1a45a55b25d2): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:44 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:44 crc kubenswrapper[4861]: E0315 00:08:44.899947 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" podUID="86b8bbef-bcb6-4395-bd33-1a45a55b25d2" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.900605 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.900693 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.900706 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.900787 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.900801 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:44Z","lastTransitionTime":"2026-03-15T00:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.907803 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.922475 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.935745 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.953855 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.969206 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:44 crc kubenswrapper[4861]: I0315 00:08:44.984767 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.001888 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.003283 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.003357 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.003383 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.003414 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.003439 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.017509 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.028389 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.043317 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.069209 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.081053 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.106878 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.106929 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.106945 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.106971 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.106990 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.108296 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.123137 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.135432 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.148637 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-t4ghv"] Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.149364 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.149577 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.162650 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.179593 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.195576 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.209349 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.209413 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.209431 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.209457 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.209479 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.212604 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.226000 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.234030 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.234217 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tbzt\" (UniqueName: \"kubernetes.io/projected/822fe19d-7cd0-44c8-b1ef-374b8a245328-kube-api-access-8tbzt\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.238588 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.259225 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.269190 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.282609 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.294845 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.312743 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.312805 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.312819 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.312842 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.312859 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.317502 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.331291 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.335114 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.335164 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tbzt\" (UniqueName: \"kubernetes.io/projected/822fe19d-7cd0-44c8-b1ef-374b8a245328-kube-api-access-8tbzt\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.335419 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.335581 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:45.835522507 +0000 UTC m=+119.147720080 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.355397 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.358773 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tbzt\" (UniqueName: \"kubernetes.io/projected/822fe19d-7cd0-44c8-b1ef-374b8a245328-kube-api-access-8tbzt\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.371816 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.385337 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.402649 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.409079 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.409211 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.409333 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.409660 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.409905 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.409914 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.410710 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.412499 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:45 crc kubenswrapper[4861]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:45 crc kubenswrapper[4861]: set -uo pipefail Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 15 00:08:45 crc kubenswrapper[4861]: HOSTS_FILE="/etc/hosts" Mar 15 00:08:45 crc kubenswrapper[4861]: TEMP_FILE="/etc/hosts.tmp" Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: # Make a temporary file with the old hosts file's attributes. Mar 15 00:08:45 crc kubenswrapper[4861]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 15 00:08:45 crc kubenswrapper[4861]: echo "Failed to preserve hosts file. Exiting." Mar 15 00:08:45 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:45 crc kubenswrapper[4861]: fi Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: while true; do Mar 15 00:08:45 crc kubenswrapper[4861]: declare -A svc_ips Mar 15 00:08:45 crc kubenswrapper[4861]: for svc in "${services[@]}"; do Mar 15 00:08:45 crc kubenswrapper[4861]: # Fetch service IP from cluster dns if present. We make several tries Mar 15 00:08:45 crc kubenswrapper[4861]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 15 00:08:45 crc kubenswrapper[4861]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 15 00:08:45 crc kubenswrapper[4861]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 15 00:08:45 crc kubenswrapper[4861]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:45 crc kubenswrapper[4861]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:45 crc kubenswrapper[4861]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 15 00:08:45 crc kubenswrapper[4861]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 15 00:08:45 crc kubenswrapper[4861]: for i in ${!cmds[*]} Mar 15 00:08:45 crc kubenswrapper[4861]: do Mar 15 00:08:45 crc kubenswrapper[4861]: ips=($(eval "${cmds[i]}")) Mar 15 00:08:45 crc kubenswrapper[4861]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 15 00:08:45 crc kubenswrapper[4861]: svc_ips["${svc}"]="${ips[@]}" Mar 15 00:08:45 crc kubenswrapper[4861]: break Mar 15 00:08:45 crc kubenswrapper[4861]: fi Mar 15 00:08:45 crc kubenswrapper[4861]: done Mar 15 00:08:45 crc kubenswrapper[4861]: done Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: # Update /etc/hosts only if we get valid service IPs Mar 15 00:08:45 crc kubenswrapper[4861]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 15 00:08:45 crc kubenswrapper[4861]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 15 00:08:45 crc kubenswrapper[4861]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 15 00:08:45 crc kubenswrapper[4861]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 15 00:08:45 crc kubenswrapper[4861]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 15 00:08:45 crc kubenswrapper[4861]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 15 00:08:45 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:45 crc kubenswrapper[4861]: continue Mar 15 00:08:45 crc kubenswrapper[4861]: fi Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: # Append resolver entries for services Mar 15 00:08:45 crc kubenswrapper[4861]: rc=0 Mar 15 00:08:45 crc kubenswrapper[4861]: for svc in "${!svc_ips[@]}"; do Mar 15 00:08:45 crc kubenswrapper[4861]: for ip in ${svc_ips[${svc}]}; do Mar 15 00:08:45 crc kubenswrapper[4861]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 15 00:08:45 crc kubenswrapper[4861]: done Mar 15 00:08:45 crc kubenswrapper[4861]: done Mar 15 00:08:45 crc kubenswrapper[4861]: if [[ $rc -ne 0 ]]; then Mar 15 00:08:45 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:45 crc kubenswrapper[4861]: continue Mar 15 00:08:45 crc kubenswrapper[4861]: fi Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: Mar 15 00:08:45 crc kubenswrapper[4861]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 15 00:08:45 crc kubenswrapper[4861]: # Replace /etc/hosts with our modified version if needed Mar 15 00:08:45 crc kubenswrapper[4861]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 15 00:08:45 crc kubenswrapper[4861]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 15 00:08:45 crc kubenswrapper[4861]: fi Mar 15 00:08:45 crc kubenswrapper[4861]: sleep 60 & wait Mar 15 00:08:45 crc kubenswrapper[4861]: unset svc_ips Mar 15 00:08:45 crc kubenswrapper[4861]: done Mar 15 00:08:45 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4nfg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-ffqzt_openshift-dns(3a70c127-00db-4104-a64d-26a45ec0b728): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:45 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.414632 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tkwmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.415978 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.416027 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.416046 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.416068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.416086 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.416704 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-ffqzt" podUID="3a70c127-00db-4104-a64d-26a45ec0b728" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.416835 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.519385 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.519462 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.519485 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.519512 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.519531 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.622363 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.622430 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.622452 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.622478 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.622496 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.725184 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.725243 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.725258 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.725283 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.725302 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.828804 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.828915 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.828934 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.828961 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.828979 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.840641 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.840858 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:45 crc kubenswrapper[4861]: E0315 00:08:45.840930 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:46.840911845 +0000 UTC m=+120.153109378 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.932114 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.932183 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.932202 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.932229 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:45 crc kubenswrapper[4861]: I0315 00:08:45.932251 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:45Z","lastTransitionTime":"2026-03-15T00:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.035483 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.035625 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.035645 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.035670 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.035689 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.138945 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.139011 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.139038 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.139068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.139092 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.242974 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.243065 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.243095 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.243129 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.243152 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.346100 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.346186 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.346210 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.346245 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.346269 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.408536 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.409222 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.411714 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:46 crc kubenswrapper[4861]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 15 00:08:46 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:46 crc kubenswrapper[4861]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 15 00:08:46 crc kubenswrapper[4861]: source /etc/kubernetes/apiserver-url.env Mar 15 00:08:46 crc kubenswrapper[4861]: else Mar 15 00:08:46 crc kubenswrapper[4861]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 15 00:08:46 crc kubenswrapper[4861]: exit 1 Mar 15 00:08:46 crc kubenswrapper[4861]: fi Mar 15 00:08:46 crc kubenswrapper[4861]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 15 00:08:46 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:46 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.412540 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:46 crc kubenswrapper[4861]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 15 00:08:46 crc kubenswrapper[4861]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 15 00:08:46 crc kubenswrapper[4861]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86x2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:46 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.412820 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.412845 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hk5r7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-rtlks_openshift-multus(a59701a7-ce44-4f3b-b03b-e01058001ac9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.414492 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.414519 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-rtlks" podUID="a59701a7-ce44-4f3b-b03b-e01058001ac9" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.452158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.452225 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.452243 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.452278 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.452299 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.556404 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.557054 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.557262 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.557440 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.557662 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.661608 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.661665 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.661687 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.661715 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.661737 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.765385 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.765441 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.765461 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.765486 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.765505 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.853618 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.853887 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:46 crc kubenswrapper[4861]: E0315 00:08:46.854501 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:48.854463405 +0000 UTC m=+122.166661138 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.873474 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.873537 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.873608 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.873648 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.873671 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.976711 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.976760 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.976769 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.976789 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:46 crc kubenswrapper[4861]: I0315 00:08:46.976800 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:46Z","lastTransitionTime":"2026-03-15T00:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.080218 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.080281 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.080295 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.080317 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.080331 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:47Z","lastTransitionTime":"2026-03-15T00:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.183471 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.183546 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.183581 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.183604 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.183619 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:47Z","lastTransitionTime":"2026-03-15T00:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.287691 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.288097 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.288243 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.288416 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.288575 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:47Z","lastTransitionTime":"2026-03-15T00:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:47 crc kubenswrapper[4861]: E0315 00:08:47.389659 4861 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.409141 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:47 crc kubenswrapper[4861]: E0315 00:08:47.409337 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.409677 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:47 crc kubenswrapper[4861]: E0315 00:08:47.409778 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.410094 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:47 crc kubenswrapper[4861]: E0315 00:08:47.410198 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.430077 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.443820 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.462644 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.477155 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.492053 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: E0315 00:08:47.496920 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.502016 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.515949 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.535452 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.552402 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.566266 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.578361 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.607897 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.618982 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.642414 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.657705 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:47 crc kubenswrapper[4861]: I0315 00:08:47.669411 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:48 crc kubenswrapper[4861]: I0315 00:08:48.409024 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.410792 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.410997 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:48 crc kubenswrapper[4861]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:48 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:48 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:48 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:48 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:48 crc kubenswrapper[4861]: fi Mar 15 00:08:48 crc kubenswrapper[4861]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 15 00:08:48 crc kubenswrapper[4861]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 15 00:08:48 crc kubenswrapper[4861]: ho_enable="--enable-hybrid-overlay" Mar 15 00:08:48 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 15 00:08:48 crc kubenswrapper[4861]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 15 00:08:48 crc kubenswrapper[4861]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 15 00:08:48 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:48 crc kubenswrapper[4861]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 15 00:08:48 crc kubenswrapper[4861]: --webhook-host=127.0.0.1 \ Mar 15 00:08:48 crc kubenswrapper[4861]: --webhook-port=9743 \ Mar 15 00:08:48 crc kubenswrapper[4861]: ${ho_enable} \ Mar 15 00:08:48 crc kubenswrapper[4861]: --enable-interconnect \ Mar 15 00:08:48 crc kubenswrapper[4861]: --disable-approver \ Mar 15 00:08:48 crc kubenswrapper[4861]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 15 00:08:48 crc kubenswrapper[4861]: --wait-for-kubernetes-api=200s \ Mar 15 00:08:48 crc kubenswrapper[4861]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 15 00:08:48 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:48 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:48 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.413012 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:48 crc kubenswrapper[4861]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 15 00:08:48 crc kubenswrapper[4861]: apiVersion: v1 Mar 15 00:08:48 crc kubenswrapper[4861]: clusters: Mar 15 00:08:48 crc kubenswrapper[4861]: - cluster: Mar 15 00:08:48 crc kubenswrapper[4861]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 15 00:08:48 crc kubenswrapper[4861]: server: https://api-int.crc.testing:6443 Mar 15 00:08:48 crc kubenswrapper[4861]: name: default-cluster Mar 15 00:08:48 crc kubenswrapper[4861]: contexts: Mar 15 00:08:48 crc kubenswrapper[4861]: - context: Mar 15 00:08:48 crc kubenswrapper[4861]: cluster: default-cluster Mar 15 00:08:48 crc kubenswrapper[4861]: namespace: default Mar 15 00:08:48 crc kubenswrapper[4861]: user: default-auth Mar 15 00:08:48 crc kubenswrapper[4861]: name: default-context Mar 15 00:08:48 crc kubenswrapper[4861]: current-context: default-context Mar 15 00:08:48 crc kubenswrapper[4861]: kind: Config Mar 15 00:08:48 crc kubenswrapper[4861]: preferences: {} Mar 15 00:08:48 crc kubenswrapper[4861]: users: Mar 15 00:08:48 crc kubenswrapper[4861]: - name: default-auth Mar 15 00:08:48 crc kubenswrapper[4861]: user: Mar 15 00:08:48 crc kubenswrapper[4861]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:48 crc kubenswrapper[4861]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 15 00:08:48 crc kubenswrapper[4861]: EOF Mar 15 00:08:48 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k452t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:48 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.413061 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:08:48 crc kubenswrapper[4861]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 15 00:08:48 crc kubenswrapper[4861]: if [[ -f "/env/_master" ]]; then Mar 15 00:08:48 crc kubenswrapper[4861]: set -o allexport Mar 15 00:08:48 crc kubenswrapper[4861]: source "/env/_master" Mar 15 00:08:48 crc kubenswrapper[4861]: set +o allexport Mar 15 00:08:48 crc kubenswrapper[4861]: fi Mar 15 00:08:48 crc kubenswrapper[4861]: Mar 15 00:08:48 crc kubenswrapper[4861]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 15 00:08:48 crc kubenswrapper[4861]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 15 00:08:48 crc kubenswrapper[4861]: --disable-webhook \ Mar 15 00:08:48 crc kubenswrapper[4861]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 15 00:08:48 crc kubenswrapper[4861]: --loglevel="${LOGLEVEL}" Mar 15 00:08:48 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 15 00:08:48 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.414238 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.415399 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:08:48 crc kubenswrapper[4861]: I0315 00:08:48.879269 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.879651 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:48 crc kubenswrapper[4861]: E0315 00:08:48.879828 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:08:52.879790421 +0000 UTC m=+126.191987994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:49 crc kubenswrapper[4861]: I0315 00:08:49.408289 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:49 crc kubenswrapper[4861]: E0315 00:08:49.408483 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:49 crc kubenswrapper[4861]: I0315 00:08:49.408987 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:49 crc kubenswrapper[4861]: E0315 00:08:49.409156 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:49 crc kubenswrapper[4861]: I0315 00:08:49.409219 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:49 crc kubenswrapper[4861]: E0315 00:08:49.409386 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:49 crc kubenswrapper[4861]: E0315 00:08:49.416365 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 15 00:08:49 crc kubenswrapper[4861]: E0315 00:08:49.417601 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 15 00:08:50 crc kubenswrapper[4861]: I0315 00:08:50.408760 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:50 crc kubenswrapper[4861]: E0315 00:08:50.408993 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:51 crc kubenswrapper[4861]: I0315 00:08:51.408793 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:51 crc kubenswrapper[4861]: I0315 00:08:51.408812 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:51 crc kubenswrapper[4861]: E0315 00:08:51.409111 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:51 crc kubenswrapper[4861]: I0315 00:08:51.408846 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:51 crc kubenswrapper[4861]: E0315 00:08:51.409222 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:51 crc kubenswrapper[4861]: E0315 00:08:51.409380 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.204299 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.204363 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.204375 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.204399 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.204413 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:52Z","lastTransitionTime":"2026-03-15T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.221456 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.228370 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.228449 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.228482 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.228501 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.228514 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:52Z","lastTransitionTime":"2026-03-15T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.244843 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.249144 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.249207 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.249220 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.249235 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.249249 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:52Z","lastTransitionTime":"2026-03-15T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.260665 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.265492 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.265614 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.265645 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.265684 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.265712 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:52Z","lastTransitionTime":"2026-03-15T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.277841 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.281988 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.282049 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.282070 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.282097 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.282115 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:08:52Z","lastTransitionTime":"2026-03-15T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.291425 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.291816 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.408736 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.408885 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.498486 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:08:52 crc kubenswrapper[4861]: I0315 00:08:52.931160 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.931401 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:52 crc kubenswrapper[4861]: E0315 00:08:52.931510 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:00.931483976 +0000 UTC m=+134.243681709 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:08:53 crc kubenswrapper[4861]: I0315 00:08:53.387195 4861 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 15 00:08:53 crc kubenswrapper[4861]: I0315 00:08:53.408298 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:53 crc kubenswrapper[4861]: E0315 00:08:53.408439 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:53 crc kubenswrapper[4861]: I0315 00:08:53.408704 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:53 crc kubenswrapper[4861]: E0315 00:08:53.408779 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:53 crc kubenswrapper[4861]: I0315 00:08:53.408918 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:53 crc kubenswrapper[4861]: E0315 00:08:53.409088 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:53 crc kubenswrapper[4861]: I0315 00:08:53.418775 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.285018 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.297785 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.306011 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.317254 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.329064 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.339969 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.355469 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.365197 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.383851 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.394810 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.408013 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:54 crc kubenswrapper[4861]: E0315 00:08:54.408163 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.411225 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.424098 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.437390 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.451381 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.464136 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.474957 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.488282 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.499638 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.928295 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tw5n7" event={"ID":"5674aba8-8b60-409d-b473-6ce136bed7d8","Type":"ContainerStarted","Data":"fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d"} Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.954013 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.971255 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.981989 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:54 crc kubenswrapper[4861]: I0315 00:08:54.998460 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.010259 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.024193 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.035904 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.049525 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.064606 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.073382 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.083185 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.093202 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.105363 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.115959 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.127347 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.151044 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.161237 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.408026 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.408102 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:55 crc kubenswrapper[4861]: I0315 00:08:55.408256 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:55 crc kubenswrapper[4861]: E0315 00:08:55.408250 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:55 crc kubenswrapper[4861]: E0315 00:08:55.408448 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:55 crc kubenswrapper[4861]: E0315 00:08:55.408706 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:56 crc kubenswrapper[4861]: I0315 00:08:56.408816 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:56 crc kubenswrapper[4861]: E0315 00:08:56.408973 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.408188 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:57 crc kubenswrapper[4861]: E0315 00:08:57.408342 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.408359 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:57 crc kubenswrapper[4861]: E0315 00:08:57.408447 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.408380 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:57 crc kubenswrapper[4861]: E0315 00:08:57.408544 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.425029 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.438790 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.450796 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.461219 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.473944 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.488666 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: E0315 00:08:57.498997 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.499824 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.508424 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.517275 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.523647 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.531355 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.544520 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.553022 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.561689 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.570934 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.579657 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.586850 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.938381 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" event={"ID":"86b8bbef-bcb6-4395-bd33-1a45a55b25d2","Type":"ContainerStarted","Data":"c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6"} Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.938460 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" event={"ID":"86b8bbef-bcb6-4395-bd33-1a45a55b25d2","Type":"ContainerStarted","Data":"ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598"} Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.940432 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401" exitCode=0 Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.940479 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401"} Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.949890 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.960167 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.969376 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.980647 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:57 crc kubenswrapper[4861]: I0315 00:08:57.992398 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.000671 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.010137 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.017200 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.026434 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.044304 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.051953 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.061291 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.072247 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.081025 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.090013 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.108268 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.122888 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.134142 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.143658 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.159810 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.170041 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.182387 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.195328 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.208369 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.216695 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.258454 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.279184 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.292122 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.299656 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.307587 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.324650 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.335909 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.348112 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.361191 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.408416 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:08:58 crc kubenswrapper[4861]: E0315 00:08:58.408549 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.946645 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2" exitCode=0 Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.946702 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2"} Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.958603 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.970071 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:58 crc kubenswrapper[4861]: I0315 00:08:58.988961 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.014043 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.025819 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.035917 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.050118 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.063782 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.073613 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.103869 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.116113 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.127669 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.140708 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.151532 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.158468 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.168159 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.176575 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.408714 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.408797 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:08:59 crc kubenswrapper[4861]: E0315 00:08:59.408826 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.408796 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:08:59 crc kubenswrapper[4861]: E0315 00:08:59.409842 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:08:59 crc kubenswrapper[4861]: E0315 00:08:59.410184 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.428629 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.953464 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" exitCode=0 Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.953528 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.957135 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b" exitCode=0 Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.957176 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b"} Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.964157 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ffqzt" event={"ID":"3a70c127-00db-4104-a64d-26a45ec0b728","Type":"ContainerStarted","Data":"3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462"} Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.972229 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.985443 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:08:59 crc kubenswrapper[4861]: I0315 00:08:59.999709 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.010328 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.025423 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.039488 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.051057 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.059354 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.069490 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.086487 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.095077 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.110734 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.117966 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.129355 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.139032 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.152110 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.161976 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.173732 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.189188 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.197825 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.210478 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.220369 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.232286 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.244267 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.254197 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.271250 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.279482 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.291707 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.304722 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.313125 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.319930 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.326501 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.353163 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.362935 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.377277 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.388905 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.408619 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:00 crc kubenswrapper[4861]: E0315 00:09:00.408795 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.963125 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:00 crc kubenswrapper[4861]: E0315 00:09:00.963252 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:00 crc kubenswrapper[4861]: E0315 00:09:00.963307 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:16.963290601 +0000 UTC m=+150.275488134 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.976236 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980444 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980522 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980534 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980545 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980577 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.980590 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.985146 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3" exitCode=0 Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.985219 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.988925 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.988987 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390"} Mar 15 00:09:00 crc kubenswrapper[4861]: I0315 00:09:00.993042 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.009734 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.026429 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.036449 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.047104 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.071419 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.083237 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.114233 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.125446 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.138414 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.152941 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.168796 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.180847 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.194090 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.208744 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.220078 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.230663 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.237398 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.252267 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.260978 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.272313 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.281654 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.298400 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.307317 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.317545 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.326922 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.335310 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.346068 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.355296 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.372189 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.382169 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.391036 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.404061 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.408461 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.408497 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.408667 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:01 crc kubenswrapper[4861]: E0315 00:09:01.408755 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:01 crc kubenswrapper[4861]: E0315 00:09:01.408965 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:01 crc kubenswrapper[4861]: E0315 00:09:01.409039 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.415727 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.426360 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:01 crc kubenswrapper[4861]: I0315 00:09:01.436634 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.000276 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630" exitCode=0 Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.000379 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630"} Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.014168 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.024897 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.036905 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.049992 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.059926 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.074468 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.087113 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.095917 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.106914 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.119824 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.141241 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.159310 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.172010 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.188233 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.197299 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.209108 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.219525 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.233369 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.408249 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.408374 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.500360 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.571374 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.571431 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.571445 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.571468 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.571483 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:02Z","lastTransitionTime":"2026-03-15T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.581360 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.585335 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.585388 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.585406 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.585433 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.585451 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:02Z","lastTransitionTime":"2026-03-15T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.603052 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.610957 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.611025 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.611040 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.611344 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.611386 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:02Z","lastTransitionTime":"2026-03-15T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.623102 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.627397 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.627428 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.627446 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.627467 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.627484 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:02Z","lastTransitionTime":"2026-03-15T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.647620 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.656271 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.656320 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.656334 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.656357 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:02 crc kubenswrapper[4861]: I0315 00:09:02.656374 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:02Z","lastTransitionTime":"2026-03-15T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.672000 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:02 crc kubenswrapper[4861]: E0315 00:09:02.672164 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.009665 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.011277 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerStarted","Data":"c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6"} Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.015532 4861 generic.go:334] "Generic (PLEG): container finished" podID="a59701a7-ce44-4f3b-b03b-e01058001ac9" containerID="033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223" exitCode=0 Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.015586 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerDied","Data":"033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223"} Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.042002 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.056974 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.069392 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.079541 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.091227 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.102276 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.112812 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.122286 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.141163 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.156021 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.172954 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.186867 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.198667 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.207676 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.234967 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.244152 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.259981 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.271715 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.296903 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.307658 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.321949 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.333760 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.348985 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.362360 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.377668 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.389611 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.399525 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.408384 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:03 crc kubenswrapper[4861]: E0315 00:09:03.408663 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.408907 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:03 crc kubenswrapper[4861]: E0315 00:09:03.408989 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.410098 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.410334 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:03 crc kubenswrapper[4861]: E0315 00:09:03.410414 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.419616 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.437095 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.451228 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.467614 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.477424 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.488407 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.512181 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:03 crc kubenswrapper[4861]: I0315 00:09:03.526887 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.021668 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9"} Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.021737 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7"} Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.028751 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" event={"ID":"a59701a7-ce44-4f3b-b03b-e01058001ac9","Type":"ContainerStarted","Data":"df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9"} Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.034750 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.048426 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.063005 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.078370 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.110144 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.119973 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.133716 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.143945 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.157225 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.167898 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.190102 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.202781 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.214829 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.230002 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.251360 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.264136 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.278383 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.297816 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.313426 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.341883 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.357800 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.375329 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.397003 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.408604 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:04 crc kubenswrapper[4861]: E0315 00:09:04.408842 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.420910 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.447134 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.481097 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.499104 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.526473 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.552330 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.578480 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.592844 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.617261 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.638312 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.656689 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.676093 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:04 crc kubenswrapper[4861]: I0315 00:09:04.692293 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:05 crc kubenswrapper[4861]: I0315 00:09:05.408681 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:05 crc kubenswrapper[4861]: E0315 00:09:05.409147 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:05 crc kubenswrapper[4861]: I0315 00:09:05.408779 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:05 crc kubenswrapper[4861]: E0315 00:09:05.409254 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:05 crc kubenswrapper[4861]: I0315 00:09:05.408779 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:05 crc kubenswrapper[4861]: E0315 00:09:05.409343 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.043473 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac"} Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.044089 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.067497 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.080847 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.086067 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.109424 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.126496 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.146012 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.163183 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.184921 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.214511 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.228408 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.243715 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.257099 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.277712 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.300801 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.326023 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.341390 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.357874 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.376884 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.392113 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.408479 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:06 crc kubenswrapper[4861]: E0315 00:09:06.408694 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.414135 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.435049 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.453317 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.472902 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.491383 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.505047 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.514473 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.532062 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.544049 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.561023 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.579472 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.591895 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.601722 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.613083 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.633220 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.643122 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.656517 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:06 crc kubenswrapper[4861]: I0315 00:09:06.673034 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.046455 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.046782 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.071992 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.087720 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.102173 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.114748 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.134637 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.146991 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.160787 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.174229 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.186659 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.198210 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.208466 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.221210 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.236484 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.249032 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.264807 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.281142 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.303974 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.322308 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.345812 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.408870 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.408921 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:07 crc kubenswrapper[4861]: E0315 00:09:07.409006 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.409085 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:07 crc kubenswrapper[4861]: E0315 00:09:07.409191 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:07 crc kubenswrapper[4861]: E0315 00:09:07.409267 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.421222 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.434807 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.447838 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.465152 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.479789 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.496000 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: E0315 00:09:07.500896 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.512730 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.525732 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.536480 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.549509 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.568846 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.585352 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.597709 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.608489 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.622651 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.633687 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.656331 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:07 crc kubenswrapper[4861]: I0315 00:09:07.666341 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.049591 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e"} Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.050912 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/0.log" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.052857 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac" exitCode=1 Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.052885 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac"} Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.053660 4861 scope.go:117] "RemoveContainer" containerID="dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.068092 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.083592 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.095057 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.109146 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.123089 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.136587 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.148464 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.175987 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.192332 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.211339 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.223925 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.235365 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.245901 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.254902 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.265114 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.283354 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.295252 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.310112 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.320778 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.337734 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.369060 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:07Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI0315 00:09:07.470739 6842 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0315 00:09:07.470745 6842 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0315 00:09:07.470763 6842 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0315 00:09:07.470771 6842 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0315 00:09:07.470777 6842 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0315 00:09:07.470784 6842 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0315 00:09:07.470815 6842 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0315 00:09:07.471491 6842 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0315 00:09:07.471838 6842 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0315 00:09:07.471888 6842 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0315 00:09:07.471926 6842 factory.go:656] Stopping watch factory\\\\nI0315 00:09:07.471942 6842 ovnkube.go:599] Stopped ovnkube\\\\nI0315 00:09:07.471971 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0315 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.399180 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.408392 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:08 crc kubenswrapper[4861]: E0315 00:09:08.408540 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.439030 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.488435 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.519982 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.558877 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.608915 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.639625 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.680082 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.719021 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.760280 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.798788 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.835671 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.881087 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.920737 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:08 crc kubenswrapper[4861]: I0315 00:09:08.958186 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.057229 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/1.log" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.057950 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/0.log" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.060827 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1" exitCode=1 Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.060865 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1"} Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.060896 4861 scope.go:117] "RemoveContainer" containerID="dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.061377 4861 scope.go:117] "RemoveContainer" containerID="113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1" Mar 15 00:09:09 crc kubenswrapper[4861]: E0315 00:09:09.061598 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.073688 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.086790 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.096546 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.116043 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.159408 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.197514 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.239015 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.278842 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.318366 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.368956 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc74f0b991fa24a0cd3183ea0142e8b060f30c88351eb3a7e9b4a3cf190a8cac\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:07Z\\\",\\\"message\\\":\\\"mers/externalversions/factory.go:140\\\\nI0315 00:09:07.470739 6842 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0315 00:09:07.470745 6842 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0315 00:09:07.470763 6842 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0315 00:09:07.470771 6842 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0315 00:09:07.470777 6842 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0315 00:09:07.470784 6842 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0315 00:09:07.470815 6842 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0315 00:09:07.471491 6842 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0315 00:09:07.471838 6842 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0315 00:09:07.471888 6842 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0315 00:09:07.471926 6842 factory.go:656] Stopping watch factory\\\\nI0315 00:09:07.471942 6842 ovnkube.go:599] Stopped ovnkube\\\\nI0315 00:09:07.471971 6842 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0315 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"message\\\":\\\"penshift-ingress-canary/ingress-canary LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8888, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0315 00:09:08.859638 6975 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0315 00:09:08.861974 6975 services_controller.go:444] Built service openshift-ingress-canary/ingress-canary LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0315 00:09:08.861973 6975 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.399337 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.407850 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.407867 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:09 crc kubenswrapper[4861]: E0315 00:09:09.407952 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.408037 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:09 crc kubenswrapper[4861]: E0315 00:09:09.408202 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:09 crc kubenswrapper[4861]: E0315 00:09:09.408415 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.446270 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.484130 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.519488 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.561630 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.606709 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.637859 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:09 crc kubenswrapper[4861]: I0315 00:09:09.678761 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.067064 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/1.log" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.070882 4861 scope.go:117] "RemoveContainer" containerID="113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1" Mar 15 00:09:10 crc kubenswrapper[4861]: E0315 00:09:10.071071 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.085197 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.103885 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.120655 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.135830 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.153728 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.169648 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.185518 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.197465 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.208836 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.228365 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"message\\\":\\\"penshift-ingress-canary/ingress-canary LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8888, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0315 00:09:08.859638 6975 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0315 00:09:08.861974 6975 services_controller.go:444] Built service openshift-ingress-canary/ingress-canary LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0315 00:09:08.861973 6975 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.239779 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.272874 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.286940 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.307647 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.322098 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.341711 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.361396 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.400865 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:10Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:10 crc kubenswrapper[4861]: I0315 00:09:10.408106 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:10 crc kubenswrapper[4861]: E0315 00:09:10.408300 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:11 crc kubenswrapper[4861]: I0315 00:09:11.408093 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:11 crc kubenswrapper[4861]: I0315 00:09:11.408189 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:11 crc kubenswrapper[4861]: E0315 00:09:11.408287 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:11 crc kubenswrapper[4861]: E0315 00:09:11.408407 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:11 crc kubenswrapper[4861]: I0315 00:09:11.408473 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:11 crc kubenswrapper[4861]: E0315 00:09:11.408693 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.408518 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.408741 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.502193 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.706636 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.706721 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.706739 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.706764 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.706784 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:12Z","lastTransitionTime":"2026-03-15T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.724651 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:12Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.729035 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.729066 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.729077 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.729092 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.729102 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:12Z","lastTransitionTime":"2026-03-15T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.741850 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:12Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.744867 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.744904 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.744915 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.744928 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.744938 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:12Z","lastTransitionTime":"2026-03-15T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.762267 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:12Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.765921 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.765951 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.765962 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.765977 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.765987 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:12Z","lastTransitionTime":"2026-03-15T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.776846 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:12Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.780716 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.780748 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.780758 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.780772 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:12 crc kubenswrapper[4861]: I0315 00:09:12.780783 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:12Z","lastTransitionTime":"2026-03-15T00:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.791734 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:12Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:12 crc kubenswrapper[4861]: E0315 00:09:12.791849 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.408205 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.408278 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.408327 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.408423 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.408720 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.408803 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.420994 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.512661 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.512859 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:10:17.512823898 +0000 UTC m=+210.825021451 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.512993 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.513038 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.513103 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.513183 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.513204 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:17.513178707 +0000 UTC m=+210.825376280 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.513223 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:17.513215779 +0000 UTC m=+210.825413312 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.614328 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:13 crc kubenswrapper[4861]: I0315 00:09:13.614421 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614647 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614646 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614725 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614750 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614667 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614829 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614833 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:17.614806236 +0000 UTC m=+210.927003809 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:09:13 crc kubenswrapper[4861]: E0315 00:09:13.614887 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:17.614870898 +0000 UTC m=+210.927068451 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:09:14 crc kubenswrapper[4861]: I0315 00:09:14.408396 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:14 crc kubenswrapper[4861]: E0315 00:09:14.408651 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:15 crc kubenswrapper[4861]: I0315 00:09:15.408166 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:15 crc kubenswrapper[4861]: I0315 00:09:15.408166 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:15 crc kubenswrapper[4861]: E0315 00:09:15.408588 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:15 crc kubenswrapper[4861]: E0315 00:09:15.408636 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:15 crc kubenswrapper[4861]: I0315 00:09:15.408188 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:15 crc kubenswrapper[4861]: E0315 00:09:15.408709 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:16 crc kubenswrapper[4861]: I0315 00:09:16.408508 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:16 crc kubenswrapper[4861]: E0315 00:09:16.408760 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.051623 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.051820 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.051924 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:09:49.051903125 +0000 UTC m=+182.364100658 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.408261 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.408308 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.408302 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.408414 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.408623 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.408865 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.421823 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.436656 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.448604 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.459965 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.475550 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.494240 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"message\\\":\\\"penshift-ingress-canary/ingress-canary LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8888, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0315 00:09:08.859638 6975 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0315 00:09:08.861974 6975 services_controller.go:444] Built service openshift-ingress-canary/ingress-canary LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0315 00:09:08.861973 6975 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: E0315 00:09:17.502589 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.504872 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.518435 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.529124 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.539160 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.548797 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.568281 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.579534 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.596470 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.618352 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.636839 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.650861 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.666785 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:17 crc kubenswrapper[4861]: I0315 00:09:17.685142 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:18 crc kubenswrapper[4861]: I0315 00:09:18.408037 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:18 crc kubenswrapper[4861]: E0315 00:09:18.408374 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:19 crc kubenswrapper[4861]: I0315 00:09:19.408683 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:19 crc kubenswrapper[4861]: E0315 00:09:19.408869 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:19 crc kubenswrapper[4861]: I0315 00:09:19.409330 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:19 crc kubenswrapper[4861]: I0315 00:09:19.409351 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:19 crc kubenswrapper[4861]: E0315 00:09:19.409467 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:19 crc kubenswrapper[4861]: E0315 00:09:19.409659 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:20 crc kubenswrapper[4861]: I0315 00:09:20.408327 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:20 crc kubenswrapper[4861]: E0315 00:09:20.408529 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:21 crc kubenswrapper[4861]: I0315 00:09:21.408783 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:21 crc kubenswrapper[4861]: E0315 00:09:21.409031 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:21 crc kubenswrapper[4861]: I0315 00:09:21.409421 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:21 crc kubenswrapper[4861]: E0315 00:09:21.409615 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:21 crc kubenswrapper[4861]: I0315 00:09:21.409879 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:21 crc kubenswrapper[4861]: E0315 00:09:21.410101 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.408518 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:22 crc kubenswrapper[4861]: E0315 00:09:22.408818 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.409957 4861 scope.go:117] "RemoveContainer" containerID="113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1" Mar 15 00:09:22 crc kubenswrapper[4861]: E0315 00:09:22.504312 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.917799 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.917879 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.917898 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.917930 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.917954 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:22Z","lastTransitionTime":"2026-03-15T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:22 crc kubenswrapper[4861]: E0315 00:09:22.945169 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:22Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.957162 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.957238 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.957262 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.957295 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.957319 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:22Z","lastTransitionTime":"2026-03-15T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:22 crc kubenswrapper[4861]: E0315 00:09:22.982366 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:22Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.987151 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.987203 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.987215 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.987230 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:22 crc kubenswrapper[4861]: I0315 00:09:22.987242 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:22Z","lastTransitionTime":"2026-03-15T00:09:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.009341 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.013034 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.013079 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.013092 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.013112 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.013124 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:23Z","lastTransitionTime":"2026-03-15T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.029929 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.034254 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.034292 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.034304 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.034327 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.034339 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:23Z","lastTransitionTime":"2026-03-15T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.049251 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.049367 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.120918 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/1.log" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.123418 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58"} Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.123935 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.137189 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.148760 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.159863 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.180876 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"message\\\":\\\"penshift-ingress-canary/ingress-canary LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8888, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0315 00:09:08.859638 6975 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0315 00:09:08.861974 6975 services_controller.go:444] Built service openshift-ingress-canary/ingress-canary LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0315 00:09:08.861973 6975 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.194319 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.212347 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.239928 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.254952 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.267432 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.296707 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.311318 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.324786 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.364523 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.394030 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.407035 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.408190 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.408239 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.408279 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.408402 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.408494 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:23 crc kubenswrapper[4861]: E0315 00:09:23.408578 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.422502 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.442396 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.459927 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:23 crc kubenswrapper[4861]: I0315 00:09:23.475217 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.128763 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/2.log" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.129890 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/1.log" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.134298 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" exitCode=1 Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.134365 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58"} Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.134437 4861 scope.go:117] "RemoveContainer" containerID="113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.135398 4861 scope.go:117] "RemoveContainer" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" Mar 15 00:09:24 crc kubenswrapper[4861]: E0315 00:09:24.135689 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.163506 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.180085 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.194674 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.207967 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.231004 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.253330 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.271947 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.289401 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.308422 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.337540 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://113329be410b297ea937edf5e39c5a71b05cd294aac0676d7a52a83c83c531f1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"message\\\":\\\"penshift-ingress-canary/ingress-canary LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.34\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8888, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0315 00:09:08.859638 6975 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0315 00:09:08.861974 6975 services_controller.go:444] Built service openshift-ingress-canary/ingress-canary LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF0315 00:09:08.861973 6975 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.354004 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.387062 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.408194 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.408728 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: E0315 00:09:24.409153 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.441930 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.459597 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.480004 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.503108 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.524015 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:24 crc kubenswrapper[4861]: I0315 00:09:24.546503 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:24Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.139139 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/2.log" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.146006 4861 scope.go:117] "RemoveContainer" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" Mar 15 00:09:25 crc kubenswrapper[4861]: E0315 00:09:25.146187 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.182872 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.204696 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.229742 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.251503 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.274808 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.301084 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.322247 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.344954 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.370023 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.389796 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.408644 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.408644 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.408834 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:25 crc kubenswrapper[4861]: E0315 00:09:25.408979 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:25 crc kubenswrapper[4861]: E0315 00:09:25.409187 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:25 crc kubenswrapper[4861]: E0315 00:09:25.409256 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.409304 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.425582 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.457229 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.474745 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.497451 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.518396 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.538647 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.556680 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:25 crc kubenswrapper[4861]: I0315 00:09:25.576127 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:26 crc kubenswrapper[4861]: I0315 00:09:26.408236 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:26 crc kubenswrapper[4861]: E0315 00:09:26.408450 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.408588 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:27 crc kubenswrapper[4861]: E0315 00:09:27.408854 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.408942 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.408951 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:27 crc kubenswrapper[4861]: E0315 00:09:27.409033 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:27 crc kubenswrapper[4861]: E0315 00:09:27.409182 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.450792 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.472184 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.503805 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: E0315 00:09:27.508281 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.524346 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.548663 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.570287 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.591458 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.615121 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.635676 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.655208 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.674081 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.693479 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.716806 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.734178 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.750533 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.766120 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.777927 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.796701 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:27 crc kubenswrapper[4861]: I0315 00:09:27.836622 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:28 crc kubenswrapper[4861]: I0315 00:09:28.408525 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:28 crc kubenswrapper[4861]: E0315 00:09:28.408863 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:29 crc kubenswrapper[4861]: I0315 00:09:29.408164 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:29 crc kubenswrapper[4861]: I0315 00:09:29.408174 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:29 crc kubenswrapper[4861]: I0315 00:09:29.408219 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:29 crc kubenswrapper[4861]: E0315 00:09:29.408538 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:29 crc kubenswrapper[4861]: E0315 00:09:29.408666 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:29 crc kubenswrapper[4861]: E0315 00:09:29.408969 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:30 crc kubenswrapper[4861]: I0315 00:09:30.407934 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:30 crc kubenswrapper[4861]: E0315 00:09:30.408183 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:31 crc kubenswrapper[4861]: I0315 00:09:31.408009 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:31 crc kubenswrapper[4861]: I0315 00:09:31.408055 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:31 crc kubenswrapper[4861]: E0315 00:09:31.408493 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:31 crc kubenswrapper[4861]: I0315 00:09:31.408055 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:31 crc kubenswrapper[4861]: E0315 00:09:31.408641 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:31 crc kubenswrapper[4861]: E0315 00:09:31.408723 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:32 crc kubenswrapper[4861]: I0315 00:09:32.408937 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:32 crc kubenswrapper[4861]: E0315 00:09:32.409192 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:32 crc kubenswrapper[4861]: E0315 00:09:32.510175 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.369133 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.369215 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.369236 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.369262 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.369280 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:33Z","lastTransitionTime":"2026-03-15T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.392424 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:33Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.399886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.399960 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.399981 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.400008 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.400029 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:33Z","lastTransitionTime":"2026-03-15T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.408506 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.408551 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.408647 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.408755 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.408849 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.408967 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.424782 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:33Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.430469 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.430527 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.430545 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.430598 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.430617 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:33Z","lastTransitionTime":"2026-03-15T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.451924 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:33Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.459040 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.459109 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.459128 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.459158 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.459178 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:33Z","lastTransitionTime":"2026-03-15T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.486243 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:33Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.492201 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.492282 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.492301 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.492337 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:33 crc kubenswrapper[4861]: I0315 00:09:33.492358 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:33Z","lastTransitionTime":"2026-03-15T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.517356 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:33Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:33 crc kubenswrapper[4861]: E0315 00:09:33.517653 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:34 crc kubenswrapper[4861]: I0315 00:09:34.408688 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:34 crc kubenswrapper[4861]: E0315 00:09:34.408812 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:35 crc kubenswrapper[4861]: I0315 00:09:35.693978 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:35 crc kubenswrapper[4861]: I0315 00:09:35.694150 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:35 crc kubenswrapper[4861]: E0315 00:09:35.695979 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:35 crc kubenswrapper[4861]: I0315 00:09:35.696311 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:35 crc kubenswrapper[4861]: E0315 00:09:35.696469 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:35 crc kubenswrapper[4861]: E0315 00:09:35.696703 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:35 crc kubenswrapper[4861]: I0315 00:09:35.697320 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:35 crc kubenswrapper[4861]: E0315 00:09:35.697735 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.408576 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.408617 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.408639 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.408665 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.408775 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.409007 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.409156 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.409454 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.410238 4861 scope.go:117] "RemoveContainer" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.410412 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.427681 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.448105 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.468160 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.491468 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: E0315 00:09:37.510886 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.533869 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.551992 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.569785 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.586508 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.610968 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.628034 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.670490 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.687456 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.701333 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.719020 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.735616 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.752915 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.770166 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.783715 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:37 crc kubenswrapper[4861]: I0315 00:09:37.800531 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:37Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:39 crc kubenswrapper[4861]: I0315 00:09:39.408471 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:39 crc kubenswrapper[4861]: I0315 00:09:39.408530 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:39 crc kubenswrapper[4861]: I0315 00:09:39.408609 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:39 crc kubenswrapper[4861]: E0315 00:09:39.408741 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:39 crc kubenswrapper[4861]: E0315 00:09:39.408863 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:39 crc kubenswrapper[4861]: I0315 00:09:39.408881 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:39 crc kubenswrapper[4861]: E0315 00:09:39.408939 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:39 crc kubenswrapper[4861]: E0315 00:09:39.409068 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:41 crc kubenswrapper[4861]: I0315 00:09:41.408507 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:41 crc kubenswrapper[4861]: I0315 00:09:41.408643 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:41 crc kubenswrapper[4861]: I0315 00:09:41.408698 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:41 crc kubenswrapper[4861]: I0315 00:09:41.408755 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:41 crc kubenswrapper[4861]: E0315 00:09:41.408658 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:41 crc kubenswrapper[4861]: E0315 00:09:41.408964 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:41 crc kubenswrapper[4861]: E0315 00:09:41.409015 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:41 crc kubenswrapper[4861]: E0315 00:09:41.409087 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:42 crc kubenswrapper[4861]: E0315 00:09:42.512155 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.408904 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.408937 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.408967 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.409104 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.410027 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.410100 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.410141 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.410162 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.625462 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.625595 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.625614 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.625655 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.625677 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:43Z","lastTransitionTime":"2026-03-15T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.645481 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.651596 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.651753 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.651822 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.651899 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.651979 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:43Z","lastTransitionTime":"2026-03-15T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.672198 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.677375 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.677525 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.677631 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.677713 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.677783 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:43Z","lastTransitionTime":"2026-03-15T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.692885 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.697667 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.697925 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.698134 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.698342 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.698750 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:43Z","lastTransitionTime":"2026-03-15T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.719780 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.725202 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.725246 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.725255 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.725274 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:43 crc kubenswrapper[4861]: I0315 00:09:43.725286 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:43Z","lastTransitionTime":"2026-03-15T00:09:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.744664 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:43Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:43 crc kubenswrapper[4861]: E0315 00:09:43.744900 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:45 crc kubenswrapper[4861]: I0315 00:09:45.408865 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:45 crc kubenswrapper[4861]: I0315 00:09:45.409017 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:45 crc kubenswrapper[4861]: I0315 00:09:45.409147 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:45 crc kubenswrapper[4861]: E0315 00:09:45.409030 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:45 crc kubenswrapper[4861]: E0315 00:09:45.409276 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:45 crc kubenswrapper[4861]: I0315 00:09:45.409173 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:45 crc kubenswrapper[4861]: E0315 00:09:45.409382 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:45 crc kubenswrapper[4861]: E0315 00:09:45.409520 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.408429 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.408518 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.408464 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.408698 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:47 crc kubenswrapper[4861]: E0315 00:09:47.408682 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:47 crc kubenswrapper[4861]: E0315 00:09:47.408813 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:47 crc kubenswrapper[4861]: E0315 00:09:47.408954 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:47 crc kubenswrapper[4861]: E0315 00:09:47.409098 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.430056 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.450447 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.465067 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.478604 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.494521 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.509609 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: E0315 00:09:47.513311 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.526317 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.546165 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.581103 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.594675 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.614582 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.628424 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.648234 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.664393 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.698646 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.718296 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.738126 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.759605 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:47 crc kubenswrapper[4861]: I0315 00:09:47.779334 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:47Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:48 crc kubenswrapper[4861]: I0315 00:09:48.410475 4861 scope.go:117] "RemoveContainer" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.131985 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.132172 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.132338 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:53.132322651 +0000 UTC m=+246.444520184 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.244321 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/2.log" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.247118 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.247484 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.249041 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/0.log" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.249112 4861 generic.go:334] "Generic (PLEG): container finished" podID="8cdc187d-d699-431a-9355-5a2268bf3821" containerID="c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6" exitCode=1 Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.249146 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerDied","Data":"c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6"} Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.249540 4861 scope.go:117] "RemoveContainer" containerID="c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.265523 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.292965 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.304595 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.320728 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.343384 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.360287 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.374271 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.398843 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.408288 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.408410 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.408431 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.408523 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.408526 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.408597 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.408731 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:49 crc kubenswrapper[4861]: E0315 00:09:49.408814 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.414829 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.430533 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.443908 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.458207 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.472938 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.486745 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.503796 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.519605 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.531759 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.545757 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.558618 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.574447 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.589939 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.605241 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.624142 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.639622 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.657969 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.670981 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.681320 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.692692 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.704636 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.718473 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.732879 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.745468 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.763196 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.774652 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.792541 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.803096 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.817233 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:49 crc kubenswrapper[4861]: I0315 00:09:49.831732 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:49Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.257665 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/3.log" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.258947 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/2.log" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.263645 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" exitCode=1 Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.263755 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.263808 4861 scope.go:117] "RemoveContainer" containerID="c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.264860 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:09:50 crc kubenswrapper[4861]: E0315 00:09:50.265208 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.267401 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/0.log" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.267482 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerStarted","Data":"671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f"} Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.296961 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.318439 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.333726 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.354442 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.369742 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.386402 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.400311 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.413056 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.430662 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.446299 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.464018 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.480430 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.500844 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.528474 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.544795 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.604805 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.623233 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.648908 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.660368 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.681019 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.694381 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.711419 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.724961 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.745647 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.762718 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.774470 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.796441 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.817713 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.836351 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.854733 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.872319 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.932427 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.955350 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.970405 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:50 crc kubenswrapper[4861]: I0315 00:09:50.982999 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.004928 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:50Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.038175 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52f90f28712486f80972b4e46b397e3db7c0db98d517e906052950ef5f91f58\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:23Z\\\",\\\"message\\\":\\\"dd Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:23Z is after 2025-08-24T17:21:41Z]\\\\nI0315 00:09:23.441757 7154 services_controller.go:434] Service openshift-marketplace/redhat-operators retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{redhat-operators openshift-marketplace 8ef79441-cef6-4ba0-a073-a7b752dbbb3e 5667 0 2025-02-23 05:23:27 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true olm.service-spec-hash:97lhyg0LJh9cnJG1O4Cl7ghtE8qwBzbCJInGtY] map[] [{operators.coreos.com/v1alpha1 CatalogSource redhat-operators e1bbbbdb-a019-4415-8578-8f8fe53276e0 0xc0007a17ad 0xc0007a17ae}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:grpc,Protocol:TCP,Port:50051,TargetPort:{0 50051 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{olm.catalogSource: redhat-operators,olm.managed: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.058147 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.272116 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/3.log" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.275767 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:09:51 crc kubenswrapper[4861]: E0315 00:09:51.275906 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.292209 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.306191 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.323192 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.337636 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.350463 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.363890 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.378188 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.405297 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.408116 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.408134 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:51 crc kubenswrapper[4861]: E0315 00:09:51.408239 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.408315 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:51 crc kubenswrapper[4861]: E0315 00:09:51.408421 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.408307 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:51 crc kubenswrapper[4861]: E0315 00:09:51.408551 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:51 crc kubenswrapper[4861]: E0315 00:09:51.408654 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.418828 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.433939 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.449128 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.463847 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.477027 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.497917 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.508821 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.522280 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.538354 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.553008 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:51 crc kubenswrapper[4861]: I0315 00:09:51.568659 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:51Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:52 crc kubenswrapper[4861]: E0315 00:09:52.514946 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:53 crc kubenswrapper[4861]: I0315 00:09:53.408028 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:53 crc kubenswrapper[4861]: I0315 00:09:53.408092 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:53 crc kubenswrapper[4861]: I0315 00:09:53.408135 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:53 crc kubenswrapper[4861]: E0315 00:09:53.408305 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:53 crc kubenswrapper[4861]: I0315 00:09:53.408334 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:53 crc kubenswrapper[4861]: E0315 00:09:53.408418 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:53 crc kubenswrapper[4861]: E0315 00:09:53.408648 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:53 crc kubenswrapper[4861]: E0315 00:09:53.408727 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.135467 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.135508 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.135518 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.135536 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.135548 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:54Z","lastTransitionTime":"2026-03-15T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.157247 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.162792 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.162850 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.162869 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.162892 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.162909 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:54Z","lastTransitionTime":"2026-03-15T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.182270 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.187953 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.188009 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.188021 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.188042 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.188055 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:54Z","lastTransitionTime":"2026-03-15T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.201696 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.206004 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.206042 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.206054 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.206070 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.206083 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:54Z","lastTransitionTime":"2026-03-15T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.225212 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.230790 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.230849 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.230877 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.230909 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:09:54 crc kubenswrapper[4861]: I0315 00:09:54.230928 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:09:54Z","lastTransitionTime":"2026-03-15T00:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.245902 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:54Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:54 crc kubenswrapper[4861]: E0315 00:09:54.246132 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:09:55 crc kubenswrapper[4861]: I0315 00:09:55.408183 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:55 crc kubenswrapper[4861]: E0315 00:09:55.408412 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:55 crc kubenswrapper[4861]: I0315 00:09:55.408810 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:55 crc kubenswrapper[4861]: I0315 00:09:55.408938 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:55 crc kubenswrapper[4861]: E0315 00:09:55.408968 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:55 crc kubenswrapper[4861]: I0315 00:09:55.409023 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:55 crc kubenswrapper[4861]: E0315 00:09:55.409057 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:55 crc kubenswrapper[4861]: E0315 00:09:55.409268 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.408891 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.408988 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:57 crc kubenswrapper[4861]: E0315 00:09:57.409094 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.409962 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:57 crc kubenswrapper[4861]: E0315 00:09:57.410491 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.410608 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:57 crc kubenswrapper[4861]: E0315 00:09:57.410787 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:57 crc kubenswrapper[4861]: E0315 00:09:57.411072 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.440050 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.462239 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.482438 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.496697 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.515762 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: E0315 00:09:57.515967 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.542607 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.564038 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.583123 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.600791 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.616384 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.634778 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.662536 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.678413 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.700966 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.718437 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.743789 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.760269 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.788681 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:57 crc kubenswrapper[4861]: I0315 00:09:57.804277 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:09:57Z is after 2025-08-24T17:21:41Z" Mar 15 00:09:59 crc kubenswrapper[4861]: I0315 00:09:59.408519 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:09:59 crc kubenswrapper[4861]: I0315 00:09:59.408628 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:09:59 crc kubenswrapper[4861]: E0315 00:09:59.408786 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:09:59 crc kubenswrapper[4861]: I0315 00:09:59.408842 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:09:59 crc kubenswrapper[4861]: E0315 00:09:59.408980 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:09:59 crc kubenswrapper[4861]: E0315 00:09:59.409108 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:09:59 crc kubenswrapper[4861]: I0315 00:09:59.409268 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:09:59 crc kubenswrapper[4861]: E0315 00:09:59.409460 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:01 crc kubenswrapper[4861]: I0315 00:10:01.408680 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:01 crc kubenswrapper[4861]: I0315 00:10:01.408755 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:01 crc kubenswrapper[4861]: I0315 00:10:01.408802 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:01 crc kubenswrapper[4861]: E0315 00:10:01.409312 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:01 crc kubenswrapper[4861]: E0315 00:10:01.409113 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:01 crc kubenswrapper[4861]: I0315 00:10:01.408680 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:01 crc kubenswrapper[4861]: E0315 00:10:01.409422 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:01 crc kubenswrapper[4861]: E0315 00:10:01.409233 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:02 crc kubenswrapper[4861]: E0315 00:10:02.517782 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:03 crc kubenswrapper[4861]: I0315 00:10:03.408957 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:03 crc kubenswrapper[4861]: I0315 00:10:03.409002 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:03 crc kubenswrapper[4861]: E0315 00:10:03.409163 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:03 crc kubenswrapper[4861]: I0315 00:10:03.409251 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:03 crc kubenswrapper[4861]: E0315 00:10:03.409677 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:03 crc kubenswrapper[4861]: E0315 00:10:03.409850 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:03 crc kubenswrapper[4861]: I0315 00:10:03.409431 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:03 crc kubenswrapper[4861]: E0315 00:10:03.410260 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.409627 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.409913 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.443154 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.443242 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.443266 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.443303 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.443327 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:04Z","lastTransitionTime":"2026-03-15T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.465842 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.471804 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.471861 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.471883 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.471906 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.471925 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:04Z","lastTransitionTime":"2026-03-15T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.495277 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.502132 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.502227 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.502254 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.502293 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.502323 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:04Z","lastTransitionTime":"2026-03-15T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.525249 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.535811 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.535886 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.535906 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.535935 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.535957 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:04Z","lastTransitionTime":"2026-03-15T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.558734 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.564858 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.564927 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.564946 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.564975 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:04 crc kubenswrapper[4861]: I0315 00:10:04.564995 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:04Z","lastTransitionTime":"2026-03-15T00:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.588687 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:04Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:04 crc kubenswrapper[4861]: E0315 00:10:04.589172 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:10:05 crc kubenswrapper[4861]: I0315 00:10:05.408002 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:05 crc kubenswrapper[4861]: I0315 00:10:05.408065 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:05 crc kubenswrapper[4861]: E0315 00:10:05.408211 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:05 crc kubenswrapper[4861]: I0315 00:10:05.408286 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:05 crc kubenswrapper[4861]: I0315 00:10:05.408037 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:05 crc kubenswrapper[4861]: E0315 00:10:05.408548 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:05 crc kubenswrapper[4861]: E0315 00:10:05.408697 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:05 crc kubenswrapper[4861]: E0315 00:10:05.408936 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.408665 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.408743 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.408756 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:07 crc kubenswrapper[4861]: E0315 00:10:07.408958 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.409003 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:07 crc kubenswrapper[4861]: E0315 00:10:07.409359 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:07 crc kubenswrapper[4861]: E0315 00:10:07.409508 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:07 crc kubenswrapper[4861]: E0315 00:10:07.409549 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.447851 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.465671 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.499961 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: E0315 00:10:07.519095 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.519761 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.552180 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.575249 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.592012 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.612065 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.633381 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.651500 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.670747 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.688295 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.708523 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.727995 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.747872 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.763202 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.781002 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.808548 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:07 crc kubenswrapper[4861]: I0315 00:10:07.823618 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:07Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:09 crc kubenswrapper[4861]: I0315 00:10:09.408847 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:09 crc kubenswrapper[4861]: I0315 00:10:09.409007 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:09 crc kubenswrapper[4861]: I0315 00:10:09.409027 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:09 crc kubenswrapper[4861]: E0315 00:10:09.409210 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:09 crc kubenswrapper[4861]: I0315 00:10:09.409307 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:09 crc kubenswrapper[4861]: E0315 00:10:09.409432 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:09 crc kubenswrapper[4861]: E0315 00:10:09.409652 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:09 crc kubenswrapper[4861]: E0315 00:10:09.409850 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:11 crc kubenswrapper[4861]: I0315 00:10:11.408505 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:11 crc kubenswrapper[4861]: I0315 00:10:11.408632 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:11 crc kubenswrapper[4861]: I0315 00:10:11.408500 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:11 crc kubenswrapper[4861]: E0315 00:10:11.408712 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:11 crc kubenswrapper[4861]: I0315 00:10:11.408512 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:11 crc kubenswrapper[4861]: E0315 00:10:11.408944 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:11 crc kubenswrapper[4861]: E0315 00:10:11.409200 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:11 crc kubenswrapper[4861]: E0315 00:10:11.409455 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:12 crc kubenswrapper[4861]: E0315 00:10:12.520706 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:13 crc kubenswrapper[4861]: I0315 00:10:13.408612 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:13 crc kubenswrapper[4861]: I0315 00:10:13.408719 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:13 crc kubenswrapper[4861]: E0315 00:10:13.408841 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:13 crc kubenswrapper[4861]: E0315 00:10:13.408919 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:13 crc kubenswrapper[4861]: I0315 00:10:13.408746 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:13 crc kubenswrapper[4861]: I0315 00:10:13.408996 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:13 crc kubenswrapper[4861]: E0315 00:10:13.409096 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:13 crc kubenswrapper[4861]: E0315 00:10:13.409253 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.753307 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.753367 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.753380 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.753403 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.753414 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:14Z","lastTransitionTime":"2026-03-15T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.771172 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:14Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.775816 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.775885 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.775919 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.775951 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.775973 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:14Z","lastTransitionTime":"2026-03-15T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.797456 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:14Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.802223 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.802271 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.802289 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.802312 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.802327 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:14Z","lastTransitionTime":"2026-03-15T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.819680 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:14Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.824456 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.824501 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.824513 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.824531 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.824541 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:14Z","lastTransitionTime":"2026-03-15T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.837081 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:14Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.841098 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.841129 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.841143 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.841160 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:14 crc kubenswrapper[4861]: I0315 00:10:14.841174 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:14Z","lastTransitionTime":"2026-03-15T00:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.859505 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:14Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:14 crc kubenswrapper[4861]: E0315 00:10:14.859680 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:10:15 crc kubenswrapper[4861]: I0315 00:10:15.408146 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:15 crc kubenswrapper[4861]: E0315 00:10:15.408338 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:15 crc kubenswrapper[4861]: I0315 00:10:15.408887 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:15 crc kubenswrapper[4861]: I0315 00:10:15.408893 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:15 crc kubenswrapper[4861]: I0315 00:10:15.408992 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:15 crc kubenswrapper[4861]: E0315 00:10:15.409247 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:15 crc kubenswrapper[4861]: E0315 00:10:15.409345 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:15 crc kubenswrapper[4861]: I0315 00:10:15.409551 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:10:15 crc kubenswrapper[4861]: E0315 00:10:15.409526 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:15 crc kubenswrapper[4861]: E0315 00:10:15.409886 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.408603 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.408715 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.408865 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.408863 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.408896 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.409074 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.409111 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.409635 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.431414 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.448633 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.466285 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.498543 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.514716 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.521461 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.539168 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.565193 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.591653 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.591924 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:12:19.59189261 +0000 UTC m=+332.904090153 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.591813 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.592125 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.592176 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.592348 4861 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.592360 4861 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.592423 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:12:19.592404054 +0000 UTC m=+332.904601597 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.592449 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:12:19.592438935 +0000 UTC m=+332.904636588 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.610399 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.640864 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.657848 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.672296 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.691405 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.692779 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.692878 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693054 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693105 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693122 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693145 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693186 4861 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693234 4861 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693355 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-15 00:12:19.693320213 +0000 UTC m=+333.005517786 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:10:17 crc kubenswrapper[4861]: E0315 00:10:17.693780 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-15 00:12:19.693749734 +0000 UTC m=+333.005947277 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.709248 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.725856 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.743394 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.757047 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.775327 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:17 crc kubenswrapper[4861]: I0315 00:10:17.793688 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:17Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:19 crc kubenswrapper[4861]: I0315 00:10:19.408203 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:19 crc kubenswrapper[4861]: I0315 00:10:19.408284 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:19 crc kubenswrapper[4861]: E0315 00:10:19.408366 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:19 crc kubenswrapper[4861]: I0315 00:10:19.408582 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:19 crc kubenswrapper[4861]: I0315 00:10:19.408630 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:19 crc kubenswrapper[4861]: E0315 00:10:19.408654 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:19 crc kubenswrapper[4861]: E0315 00:10:19.408730 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:19 crc kubenswrapper[4861]: E0315 00:10:19.408795 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:21 crc kubenswrapper[4861]: I0315 00:10:21.408519 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:21 crc kubenswrapper[4861]: I0315 00:10:21.408613 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:21 crc kubenswrapper[4861]: E0315 00:10:21.408791 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:21 crc kubenswrapper[4861]: I0315 00:10:21.408822 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:21 crc kubenswrapper[4861]: E0315 00:10:21.408953 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:21 crc kubenswrapper[4861]: E0315 00:10:21.409031 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:21 crc kubenswrapper[4861]: I0315 00:10:21.409707 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:21 crc kubenswrapper[4861]: E0315 00:10:21.409855 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:22 crc kubenswrapper[4861]: E0315 00:10:22.522592 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:23 crc kubenswrapper[4861]: I0315 00:10:23.408285 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:23 crc kubenswrapper[4861]: I0315 00:10:23.408350 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:23 crc kubenswrapper[4861]: I0315 00:10:23.408416 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:23 crc kubenswrapper[4861]: I0315 00:10:23.408484 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:23 crc kubenswrapper[4861]: E0315 00:10:23.408479 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:23 crc kubenswrapper[4861]: E0315 00:10:23.408682 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:23 crc kubenswrapper[4861]: E0315 00:10:23.408772 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:23 crc kubenswrapper[4861]: E0315 00:10:23.408858 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.145552 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.145660 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.145677 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.145702 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.145721 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:25Z","lastTransitionTime":"2026-03-15T00:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.168212 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.174754 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.174820 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.174840 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.174866 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.174886 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:25Z","lastTransitionTime":"2026-03-15T00:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.196405 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.202319 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.202387 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.202402 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.202422 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.202437 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:25Z","lastTransitionTime":"2026-03-15T00:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.231395 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.237783 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.238049 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.238135 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.238223 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.238298 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:25Z","lastTransitionTime":"2026-03-15T00:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.260089 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.265360 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.265413 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.265431 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.265457 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.265476 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:25Z","lastTransitionTime":"2026-03-15T00:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.283274 4861 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-15T00:10:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"963bfba1-4eec-4efc-bcce-77026d8b2488\\\",\\\"systemUUID\\\":\\\"0787823d-1679-41ae-95f7-6ed423e90e04\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:25Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.283548 4861 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.408683 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.408764 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.408679 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.408867 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:25 crc kubenswrapper[4861]: I0315 00:10:25.408776 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.408997 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.409079 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:25 crc kubenswrapper[4861]: E0315 00:10:25.409267 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:26 crc kubenswrapper[4861]: I0315 00:10:26.409914 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:10:26 crc kubenswrapper[4861]: E0315 00:10:26.410122 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.407860 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.407969 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:27 crc kubenswrapper[4861]: E0315 00:10:27.408103 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:27 crc kubenswrapper[4861]: E0315 00:10:27.408359 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.408941 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:27 crc kubenswrapper[4861]: E0315 00:10:27.409153 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.409321 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:27 crc kubenswrapper[4861]: E0315 00:10:27.409762 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.425128 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86b8bbef-bcb6-4395-bd33-1a45a55b25d2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef1ff312277a09d85ba8dbd7698154d6685801d5125b9442e10675f9a2c25598\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6ba04587a4437a0aa1af8735cde4dc9b0517b15ea6dc82d843d84734536aad6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v7bf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-r25wl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.448971 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"288c1ed8-4ba5-4c18-8a81-41bc2c6d1cea\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://909f72288d2ecbaa6001d8e6730e769a0ba1e21d7b3910738bc0395c278649bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fafd76528faac92787f5a85f837c59dfc57cbbfdfde9e878a6ff8a9b1ec2ce01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f99940fbda8f55556820150c2c93e931370be5d887e5acc143866a496466ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3fb463dcfb7e2a31c2176a17e4ae525e68b3b608301467e6259019d2e4c0ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a0cb7c97131c69612d65b3ab9a5d3c0d2d2b1454bb3c38371a4f6e9e75749a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d720add4b3b06f0d3b6085e3d2b1e00599d4d56bb1674ef750ad3a4a46c98605\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57d5660139a2ee4b8a3e721823b4a8372b47934b8f781fa5854bd64256084b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5467a183236c705b382b349f2c6de6da1bc304551afefad9ec4596989f4eb039\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.465007 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e73a7f7-b357-463e-bfd0-89b2949b9444\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e441176a8080d2e7f528b6189c111a5933ee7586bb48839486745a491c4bbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d75b5a63f18f445ed8ecd2f7dc18beb4e4b055c49100f7eba98be4900ebf55\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.492391 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rtlks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a59701a7-ce44-4f3b-b03b-e01058001ac9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df70b892fbc8664582bb297726a518b8402ab1fcb85a418d40ad23c31b7b49d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07cebb471db39a1b157814a5308d9c62977910a83bda885136ddab5ab59e8401\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2cc8888dbb55f763286691a064f3c8e49e2be9f98ef1c23009dcb106b8e92b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ca01020b611658189b13b3bc57e26251fe20aa5c2213927d5c3b07cb149215b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65815d6ae1d2c3d4ec6a3c8d2baf77ec56f59182479d329495c512d16483b7b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://675a0d74d248c7dc1107f0c745c7564fb7db622631fb8a73cda527c238da3630\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://033cc72029adeab990f4ef90e568456acb6a729aac63875863544393f3620223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:09:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5r7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rtlks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.515734 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bbzq6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8cdc187d-d699-431a-9355-5a2268bf3821\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:48Z\\\",\\\"message\\\":\\\"2026-03-15T00:09:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52\\\\n2026-03-15T00:09:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d173f6ee-208d-4690-950f-b9aa5fdbde52 to /host/opt/cni/bin/\\\\n2026-03-15T00:09:03Z [verbose] multus-daemon started\\\\n2026-03-15T00:09:03Z [verbose] Readiness Indicator file check\\\\n2026-03-15T00:09:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-86x2p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bbzq6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: E0315 00:10:27.523128 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.544178 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c527691d-7c1c-4780-8a7d-ff270650a7eb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0315 00:07:52.856309 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0315 00:07:52.856456 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0315 00:07:52.857224 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3177938706/tls.crt::/tmp/serving-cert-3177938706/tls.key\\\\\\\"\\\\nI0315 00:07:53.043100 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0315 00:07:53.047279 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0315 00:07:53.047296 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0315 00:07:53.047321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0315 00:07:53.047327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0315 00:07:53.055488 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0315 00:07:53.055519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055526 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0315 00:07:53.055533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0315 00:07:53.055537 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0315 00:07:53.055542 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0315 00:07:53.055546 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0315 00:07:53.055569 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0315 00:07:53.055791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.561245 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a2679d0d-772a-4767-8148-6bfd8801057b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e5eeb4e3c0d082e1bf802105a8e7afa2035273d3c006310b130868e9799ac4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf339baacb614467b4f4cbde670f08e77912051fcec67c59d29ec1321fe178fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-15T00:07:46Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0315 00:07:14.336160 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0315 00:07:14.337848 1 observer_polling.go:159] Starting file observer\\\\nI0315 00:07:14.339413 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0315 00:07:14.340479 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0315 00:07:41.116324 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nF0315 00:07:45.102072 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:07:14Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:07:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53a9b3536384eb22f829aa475c57e02880bc49d06f7bdeae6581981d01c2b955\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93460f3eeeabbd917cb9eed43f6d7c71d96b87d20a8dc352e89225d3dfcbe826\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.572982 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5cfa9c2-80b7-4791-85e7-2ebe4135c5b2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:07:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3b5a9d2cacfb750be7983a6ffac86d5f1f310b081aa9630d694e211d9a61ef1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac2aa07d71bfe65b6553a5e687e27f66a35652e65e2c0eaaaec2fbf82ddda3d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2541324fe397be4ff25c0a326db0c34d5f8c71e334c817c64c18e6856fb4078\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d6ef327d6a25b683dd9ae7cbf81b88ec3eeac50f13d33eb32d8caee4dcad611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:06:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:06:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.582743 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"822fe19d-7cd0-44c8-b1ef-374b8a245328\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8tbzt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-t4ghv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.596206 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c82bd5ccc017002d8b3bb56c6a735d4c3373547e3e4dd6f8f4bb768cf07cb342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.614547 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.633905 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.652661 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ffqzt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a70c127-00db-4104-a64d-26a45ec0b728\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c602d473274a592ef6867265f9732d7df8102bac6cda0ee11287fbb31c58462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4nfg5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ffqzt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.665406 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5442c197-f16b-462d-b7ef-f509c8b1e858\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://782cc7699b675908912714662f7d1574e18b33bfdd00c62c8314d407bbae063f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tkwmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8m45\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.699141 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa2d91cf-7d1a-4567-9593-654417784ef1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-15T00:09:49Z\\\",\\\"message\\\":\\\"lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.219\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0315 00:09:49.341066 7396 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, buil\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-15T00:09:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pmf25_openshift-ovn-kubernetes(fa2d91cf-7d1a-4567-9593-654417784ef1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-15T00:08:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-15T00:08:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k452t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pmf25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.714989 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tw5n7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5674aba8-8b60-409d-b473-6ce136bed7d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc710fa3c5421217f7c8a66d6f99cb12e752e015e610b4da22aa4d591bfefb3d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swv5m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-15T00:08:38Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tw5n7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.735137 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.754469 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d36097dffc53fe80536ec2e75768f2746f6c4774d77cb7594376de84139f6aa9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19dde4eb731c2603858195c530cd073093a450825ddef4f156bbfa61f09267f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:27 crc kubenswrapper[4861]: I0315 00:10:27.775494 4861 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:08:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-15T00:09:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aba601c75e85865c28e7f857aeb56a32e29fb35c69ab5befdbcd197724bf599e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-15T00:09:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-15T00:10:27Z is after 2025-08-24T17:21:41Z" Mar 15 00:10:29 crc kubenswrapper[4861]: I0315 00:10:29.407916 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:29 crc kubenswrapper[4861]: E0315 00:10:29.408942 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:29 crc kubenswrapper[4861]: I0315 00:10:29.408142 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:29 crc kubenswrapper[4861]: E0315 00:10:29.409121 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:29 crc kubenswrapper[4861]: I0315 00:10:29.408172 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:29 crc kubenswrapper[4861]: E0315 00:10:29.409255 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:29 crc kubenswrapper[4861]: I0315 00:10:29.408066 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:29 crc kubenswrapper[4861]: E0315 00:10:29.409404 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:31 crc kubenswrapper[4861]: I0315 00:10:31.409015 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:31 crc kubenswrapper[4861]: I0315 00:10:31.409104 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:31 crc kubenswrapper[4861]: I0315 00:10:31.409139 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:31 crc kubenswrapper[4861]: I0315 00:10:31.409137 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:31 crc kubenswrapper[4861]: E0315 00:10:31.409220 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:31 crc kubenswrapper[4861]: E0315 00:10:31.409355 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:31 crc kubenswrapper[4861]: E0315 00:10:31.409461 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:31 crc kubenswrapper[4861]: E0315 00:10:31.409536 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:32 crc kubenswrapper[4861]: E0315 00:10:32.525446 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:33 crc kubenswrapper[4861]: I0315 00:10:33.408502 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:33 crc kubenswrapper[4861]: I0315 00:10:33.408657 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:33 crc kubenswrapper[4861]: I0315 00:10:33.408665 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:33 crc kubenswrapper[4861]: I0315 00:10:33.408779 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:33 crc kubenswrapper[4861]: E0315 00:10:33.408790 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:33 crc kubenswrapper[4861]: E0315 00:10:33.408923 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:33 crc kubenswrapper[4861]: E0315 00:10:33.409017 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:33 crc kubenswrapper[4861]: E0315 00:10:33.409086 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.408453 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:35 crc kubenswrapper[4861]: E0315 00:10:35.410022 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.408710 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:35 crc kubenswrapper[4861]: E0315 00:10:35.410320 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.408940 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:35 crc kubenswrapper[4861]: E0315 00:10:35.410453 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.408636 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:35 crc kubenswrapper[4861]: E0315 00:10:35.410609 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.443897 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/1.log" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.444640 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/0.log" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.444706 4861 generic.go:334] "Generic (PLEG): container finished" podID="8cdc187d-d699-431a-9355-5a2268bf3821" containerID="671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f" exitCode=1 Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.444754 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerDied","Data":"671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f"} Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.444805 4861 scope.go:117] "RemoveContainer" containerID="c35e310f6ce098aaad3cf36bc1f920e5b79fd41d1706026825b528b6acd9f6b6" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.445376 4861 scope.go:117] "RemoveContainer" containerID="671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f" Mar 15 00:10:35 crc kubenswrapper[4861]: E0315 00:10:35.445686 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821)\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.503226 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=118.503201064 podStartE2EDuration="1m58.503201064s" podCreationTimestamp="2026-03-15 00:08:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.501993511 +0000 UTC m=+228.814191084" watchObservedRunningTime="2026-03-15 00:10:35.503201064 +0000 UTC m=+228.815398607" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.515651 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=96.515619827 podStartE2EDuration="1m36.515619827s" podCreationTimestamp="2026-03-15 00:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.515093252 +0000 UTC m=+228.827290795" watchObservedRunningTime="2026-03-15 00:10:35.515619827 +0000 UTC m=+228.827817390" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.540507 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rtlks" podStartSLOduration=166.540480744 podStartE2EDuration="2m46.540480744s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.540147655 +0000 UTC m=+228.852345198" watchObservedRunningTime="2026-03-15 00:10:35.540480744 +0000 UTC m=+228.852678277" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.556836 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-r25wl" podStartSLOduration=165.556804512 podStartE2EDuration="2m45.556804512s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.556229977 +0000 UTC m=+228.868427520" watchObservedRunningTime="2026-03-15 00:10:35.556804512 +0000 UTC m=+228.869002075" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.575602 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=135.575574346 podStartE2EDuration="2m15.575574346s" podCreationTimestamp="2026-03-15 00:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.575506024 +0000 UTC m=+228.887703567" watchObservedRunningTime="2026-03-15 00:10:35.575574346 +0000 UTC m=+228.887771889" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.612842 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=82.612820126 podStartE2EDuration="1m22.612820126s" podCreationTimestamp="2026-03-15 00:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.593075596 +0000 UTC m=+228.905273139" watchObservedRunningTime="2026-03-15 00:10:35.612820126 +0000 UTC m=+228.925017669" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.630420 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=102.630391958 podStartE2EDuration="1m42.630391958s" podCreationTimestamp="2026-03-15 00:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.613293138 +0000 UTC m=+228.925490681" watchObservedRunningTime="2026-03-15 00:10:35.630391958 +0000 UTC m=+228.942589501" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.648068 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.648123 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.648140 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.648161 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.648178 4861 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-15T00:10:35Z","lastTransitionTime":"2026-03-15T00:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.703503 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj"] Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.704073 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.706231 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.706537 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.706657 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.707137 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.779648 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.800087 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ffqzt" podStartSLOduration=166.800065203 podStartE2EDuration="2m46.800065203s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.799285391 +0000 UTC m=+229.111482924" watchObservedRunningTime="2026-03-15 00:10:35.800065203 +0000 UTC m=+229.112262736" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.802535 4861 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817413 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0ab00f56-b57d-4781-8ba2-4ef4beae4524-service-ca\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817453 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817525 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ab00f56-b57d-4781-8ba2-4ef4beae4524-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817598 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ab00f56-b57d-4781-8ba2-4ef4beae4524-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817550 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podStartSLOduration=166.817518911 podStartE2EDuration="2m46.817518911s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.816801282 +0000 UTC m=+229.128998815" watchObservedRunningTime="2026-03-15 00:10:35.817518911 +0000 UTC m=+229.129716444" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.817785 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.852514 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tw5n7" podStartSLOduration=166.8524922 podStartE2EDuration="2m46.8524922s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:35.850487956 +0000 UTC m=+229.162685479" watchObservedRunningTime="2026-03-15 00:10:35.8524922 +0000 UTC m=+229.164689733" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.918932 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ab00f56-b57d-4781-8ba2-4ef4beae4524-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919032 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919096 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0ab00f56-b57d-4781-8ba2-4ef4beae4524-service-ca\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919131 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919173 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ab00f56-b57d-4781-8ba2-4ef4beae4524-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919214 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.919418 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0ab00f56-b57d-4781-8ba2-4ef4beae4524-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.920936 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0ab00f56-b57d-4781-8ba2-4ef4beae4524-service-ca\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.925326 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0ab00f56-b57d-4781-8ba2-4ef4beae4524-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:35 crc kubenswrapper[4861]: I0315 00:10:35.938871 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0ab00f56-b57d-4781-8ba2-4ef4beae4524-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-c4dmj\" (UID: \"0ab00f56-b57d-4781-8ba2-4ef4beae4524\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:36 crc kubenswrapper[4861]: I0315 00:10:36.020584 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" Mar 15 00:10:36 crc kubenswrapper[4861]: I0315 00:10:36.451896 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/1.log" Mar 15 00:10:36 crc kubenswrapper[4861]: I0315 00:10:36.454677 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" event={"ID":"0ab00f56-b57d-4781-8ba2-4ef4beae4524","Type":"ContainerStarted","Data":"b3ff4a820275b8a28f75e9e3a58377eabd58b00a2eb0ebe37e8c528ceaae9bcc"} Mar 15 00:10:36 crc kubenswrapper[4861]: I0315 00:10:36.454753 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" event={"ID":"0ab00f56-b57d-4781-8ba2-4ef4beae4524","Type":"ContainerStarted","Data":"a06cf792718f751a075cae69788034ddc6a8036c4a63738586f80256cf453b54"} Mar 15 00:10:36 crc kubenswrapper[4861]: I0315 00:10:36.484306 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-c4dmj" podStartSLOduration=167.48428244 podStartE2EDuration="2m47.48428244s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:36.482169053 +0000 UTC m=+229.794366616" watchObservedRunningTime="2026-03-15 00:10:36.48428244 +0000 UTC m=+229.796479973" Mar 15 00:10:37 crc kubenswrapper[4861]: I0315 00:10:37.408010 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:37 crc kubenswrapper[4861]: I0315 00:10:37.408048 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:37 crc kubenswrapper[4861]: I0315 00:10:37.408011 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:37 crc kubenswrapper[4861]: I0315 00:10:37.408102 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:37 crc kubenswrapper[4861]: E0315 00:10:37.409790 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:37 crc kubenswrapper[4861]: E0315 00:10:37.409861 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:37 crc kubenswrapper[4861]: E0315 00:10:37.409945 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:37 crc kubenswrapper[4861]: E0315 00:10:37.410215 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:37 crc kubenswrapper[4861]: E0315 00:10:37.526417 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:38 crc kubenswrapper[4861]: I0315 00:10:38.409349 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.411493 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:39 crc kubenswrapper[4861]: E0315 00:10:39.411965 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.412217 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:39 crc kubenswrapper[4861]: E0315 00:10:39.412305 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.412456 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:39 crc kubenswrapper[4861]: E0315 00:10:39.412520 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.412687 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:39 crc kubenswrapper[4861]: E0315 00:10:39.412745 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.469468 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/3.log" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.473458 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerStarted","Data":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.474192 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.515944 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podStartSLOduration=170.515920964 podStartE2EDuration="2m50.515920964s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:39.5146521 +0000 UTC m=+232.826849643" watchObservedRunningTime="2026-03-15 00:10:39.515920964 +0000 UTC m=+232.828118497" Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.516260 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t4ghv"] Mar 15 00:10:39 crc kubenswrapper[4861]: I0315 00:10:39.516364 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:39 crc kubenswrapper[4861]: E0315 00:10:39.516462 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:41 crc kubenswrapper[4861]: I0315 00:10:41.407997 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:41 crc kubenswrapper[4861]: I0315 00:10:41.408090 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:41 crc kubenswrapper[4861]: I0315 00:10:41.408157 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:41 crc kubenswrapper[4861]: I0315 00:10:41.408209 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:41 crc kubenswrapper[4861]: E0315 00:10:41.408227 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:41 crc kubenswrapper[4861]: E0315 00:10:41.408402 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:41 crc kubenswrapper[4861]: E0315 00:10:41.408541 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:41 crc kubenswrapper[4861]: E0315 00:10:41.408830 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:42 crc kubenswrapper[4861]: E0315 00:10:42.528210 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:43 crc kubenswrapper[4861]: I0315 00:10:43.408045 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:43 crc kubenswrapper[4861]: I0315 00:10:43.408160 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:43 crc kubenswrapper[4861]: I0315 00:10:43.408169 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:43 crc kubenswrapper[4861]: E0315 00:10:43.408252 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:43 crc kubenswrapper[4861]: I0315 00:10:43.408322 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:43 crc kubenswrapper[4861]: E0315 00:10:43.408510 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:43 crc kubenswrapper[4861]: E0315 00:10:43.408885 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:43 crc kubenswrapper[4861]: E0315 00:10:43.409039 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:45 crc kubenswrapper[4861]: I0315 00:10:45.408421 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:45 crc kubenswrapper[4861]: I0315 00:10:45.408513 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:45 crc kubenswrapper[4861]: I0315 00:10:45.408505 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:45 crc kubenswrapper[4861]: E0315 00:10:45.408678 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:45 crc kubenswrapper[4861]: I0315 00:10:45.408749 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:45 crc kubenswrapper[4861]: E0315 00:10:45.408944 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:45 crc kubenswrapper[4861]: E0315 00:10:45.409067 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:45 crc kubenswrapper[4861]: E0315 00:10:45.409204 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:47 crc kubenswrapper[4861]: I0315 00:10:47.408505 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:47 crc kubenswrapper[4861]: I0315 00:10:47.408505 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:47 crc kubenswrapper[4861]: E0315 00:10:47.410145 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:47 crc kubenswrapper[4861]: I0315 00:10:47.410206 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:47 crc kubenswrapper[4861]: I0315 00:10:47.410310 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:47 crc kubenswrapper[4861]: E0315 00:10:47.410444 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:47 crc kubenswrapper[4861]: E0315 00:10:47.410649 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:47 crc kubenswrapper[4861]: E0315 00:10:47.410788 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:47 crc kubenswrapper[4861]: E0315 00:10:47.529083 4861 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:10:49 crc kubenswrapper[4861]: I0315 00:10:49.408587 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:49 crc kubenswrapper[4861]: I0315 00:10:49.408666 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:49 crc kubenswrapper[4861]: I0315 00:10:49.408694 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:49 crc kubenswrapper[4861]: I0315 00:10:49.408705 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:49 crc kubenswrapper[4861]: I0315 00:10:49.409106 4861 scope.go:117] "RemoveContainer" containerID="671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f" Mar 15 00:10:49 crc kubenswrapper[4861]: E0315 00:10:49.409386 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:49 crc kubenswrapper[4861]: E0315 00:10:49.409735 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:49 crc kubenswrapper[4861]: E0315 00:10:49.409852 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:49 crc kubenswrapper[4861]: E0315 00:10:49.409929 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:50 crc kubenswrapper[4861]: I0315 00:10:50.523695 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/1.log" Mar 15 00:10:50 crc kubenswrapper[4861]: I0315 00:10:50.524313 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerStarted","Data":"4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746"} Mar 15 00:10:50 crc kubenswrapper[4861]: I0315 00:10:50.547698 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-bbzq6" podStartSLOduration=181.547669597 podStartE2EDuration="3m1.547669597s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:10:50.547531843 +0000 UTC m=+243.859729446" watchObservedRunningTime="2026-03-15 00:10:50.547669597 +0000 UTC m=+243.859867160" Mar 15 00:10:51 crc kubenswrapper[4861]: I0315 00:10:51.498852 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:51 crc kubenswrapper[4861]: E0315 00:10:51.499010 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 15 00:10:51 crc kubenswrapper[4861]: I0315 00:10:51.499585 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:51 crc kubenswrapper[4861]: E0315 00:10:51.499915 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 15 00:10:51 crc kubenswrapper[4861]: I0315 00:10:51.499959 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:51 crc kubenswrapper[4861]: E0315 00:10:51.500108 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 15 00:10:51 crc kubenswrapper[4861]: I0315 00:10:51.500458 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:51 crc kubenswrapper[4861]: E0315 00:10:51.500603 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-t4ghv" podUID="822fe19d-7cd0-44c8-b1ef-374b8a245328" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.220739 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:53 crc kubenswrapper[4861]: E0315 00:10:53.220973 4861 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:10:53 crc kubenswrapper[4861]: E0315 00:10:53.221077 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs podName:822fe19d-7cd0-44c8-b1ef-374b8a245328 nodeName:}" failed. No retries permitted until 2026-03-15 00:12:55.221055623 +0000 UTC m=+368.533253156 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs") pod "network-metrics-daemon-t4ghv" (UID: "822fe19d-7cd0-44c8-b1ef-374b8a245328") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.408868 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.408945 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.408992 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.409066 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.412106 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.412197 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.412742 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.412900 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.412975 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 15 00:10:53 crc kubenswrapper[4861]: I0315 00:10:53.414115 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.944391 4861 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.995595 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx"] Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.996483 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.996615 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d"] Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.997409 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.997818 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j2krj"] Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.998248 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.998957 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j"] Mar 15 00:10:55 crc kubenswrapper[4861]: I0315 00:10:55.999406 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.000101 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xzvqd"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.000352 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.000981 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lvzfq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.001283 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.001950 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.002483 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006309 4861 reflector.go:561] object-"openshift-cluster-samples-operator"/"samples-operator-tls": failed to list *v1.Secret: secrets "samples-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006353 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"samples-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006414 4861 reflector.go:561] object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006430 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006506 4861 reflector.go:561] object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w": failed to list *v1.Secret: secrets "cluster-samples-operator-dockercfg-xpp9w" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006541 4861 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006575 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006579 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-xpp9w\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-samples-operator-dockercfg-xpp9w\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006516 4861 reflector.go:561] object-"openshift-cluster-samples-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006607 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-login": failed to list *v1.Secret: secrets "v4-0-config-user-template-login" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006620 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-login\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006618 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006666 4861 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert": failed to list *v1.Secret: secrets "openshift-apiserver-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006679 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006699 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-provider-selection": failed to list *v1.Secret: secrets "v4-0-config-user-template-provider-selection" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006712 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-provider-selection\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006718 4861 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config": failed to list *v1.ConfigMap: configmaps "openshift-apiserver-operator-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006737 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-apiserver-operator-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006751 4861 reflector.go:561] object-"openshift-apiserver-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006763 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006774 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template": failed to list *v1.Secret: secrets "v4-0-config-system-ocp-branding-template" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006786 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-ocp-branding-template\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006843 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006855 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006891 4861 reflector.go:561] object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc": failed to list *v1.Secret: secrets "oauth-openshift-dockercfg-znhcc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006902 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-znhcc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-openshift-dockercfg-znhcc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.006952 4861 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv": failed to list *v1.Secret: secrets "openshift-apiserver-operator-dockercfg-xtcjv" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.006965 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-xtcjv\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-operator-dockercfg-xtcjv\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.007002 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-session": failed to list *v1.Secret: secrets "v4-0-config-system-session" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.007013 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-session\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-session\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.007045 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data": failed to list *v1.Secret: secrets "v4-0-config-user-idp-0-file-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.007059 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-idp-0-file-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.007099 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-router-certs": failed to list *v1.Secret: secrets "v4-0-config-system-router-certs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.007110 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-router-certs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.007156 4861 reflector.go:561] object-"openshift-authentication"/"audit": failed to list *v1.ConfigMap: configmaps "audit" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.007168 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.007241 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-serving-cert": failed to list *v1.Secret: secrets "v4-0-config-system-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.007266 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.010457 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.011021 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zw4vw"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.011187 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.011467 4861 reflector.go:561] object-"openshift-authentication"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.011498 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.012180 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4c7wr"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.012403 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lns4z"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.012801 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.013031 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbnkg"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.013113 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.013385 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.014165 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.018571 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-cliconfig": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-cliconfig" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.018651 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-cliconfig\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.028282 4861 reflector.go:561] object-"openshift-authentication"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.028341 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.028441 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-error": failed to list *v1.Secret: secrets "v4-0-config-user-template-error" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.028456 4861 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-service-ca": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.028504 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-error\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.028549 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.028712 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.029012 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.029413 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.030183 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.030332 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.030371 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.030449 4861 reflector.go:561] object-"openshift-console"/"console-config": failed to list *v1.ConfigMap: configmaps "console-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.030467 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"console-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.030546 4861 reflector.go:561] object-"openshift-console"/"oauth-serving-cert": failed to list *v1.ConfigMap: configmaps "oauth-serving-cert" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.030582 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"oauth-serving-cert\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"oauth-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.031109 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.031916 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.031992 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29558880-xw69x"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.032736 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.048028 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.049516 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.049847 4861 reflector.go:561] object-"openshift-dns-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.049930 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.049958 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.050447 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051038 4861 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.051102 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.051114 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051202 4861 reflector.go:561] object-"openshift-console"/"console-serving-cert": failed to list *v1.Secret: secrets "console-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051254 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051400 4861 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051421 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.051677 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051697 4861 reflector.go:561] object-"openshift-console"/"console-dockercfg-f62pw": failed to list *v1.Secret: secrets "console-dockercfg-f62pw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051726 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-dockercfg-f62pw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-dockercfg-f62pw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051825 4861 reflector.go:561] object-"openshift-console"/"console-oauth-config": failed to list *v1.Secret: secrets "console-oauth-config" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.051894 4861 reflector.go:561] object-"openshift-console"/"service-ca": failed to list *v1.ConfigMap: configmaps "service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051928 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051886 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"console-oauth-config\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"console-oauth-config\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.052104 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.052233 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.052403 4861 reflector.go:561] object-"openshift-dns-operator"/"metrics-tls": failed to list *v1.Secret: secrets "metrics-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.052442 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns-operator\"/\"metrics-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metrics-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.052402 4861 reflector.go:561] object-"openshift-console"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-console": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.052464 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.052477 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-console\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-console\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.052611 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.052873 4861 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.052999 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.054234 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.054365 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.054504 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.051096 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.054661 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057102 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057298 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.057379 4861 reflector.go:561] object-"openshift-dns-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns-operator": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.057413 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057437 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057567 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057610 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.057672 4861 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.057688 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057729 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057793 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057730 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057929 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057981 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.057993 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058016 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058105 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058121 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.058157 4861 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.058918 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058173 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058194 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058230 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058251 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: W0315 00:10:56.058312 4861 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Mar 15 00:10:56 crc kubenswrapper[4861]: E0315 00:10:56.059291 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058318 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.059321 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058355 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058407 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.059489 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058419 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058504 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.058597 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.062741 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.066025 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.066181 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.070744 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lgbxx"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.071670 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.071925 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.074975 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.076671 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.078194 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.078970 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.079144 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.079266 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.079352 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.080164 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.081647 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-phl6h"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.082341 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.086784 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.087263 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.088691 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.089236 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k9d4t"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.089587 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.089893 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090058 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090202 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090269 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090429 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090501 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090737 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090799 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090949 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.090982 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.091182 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.091279 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.091315 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.091515 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.092365 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095209 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095256 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095287 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-client\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095352 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095379 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-auth-proxy-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095406 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095441 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095470 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095498 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drcws\" (UniqueName: \"kubernetes.io/projected/b0c9069d-5138-4fd9-9c2f-78b5224287e8-kube-api-access-drcws\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095546 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44h8w\" (UniqueName: \"kubernetes.io/projected/2ec532f5-c27d-4a38-900c-fbbc5676a5cd-kube-api-access-44h8w\") pod \"downloads-7954f5f757-xzvqd\" (UID: \"2ec532f5-c27d-4a38-900c-fbbc5676a5cd\") " pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095595 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095623 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit-dir\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095647 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095672 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095710 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqhsv\" (UniqueName: \"kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095744 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-client\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095777 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095805 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfl8\" (UniqueName: \"kubernetes.io/projected/6ef26501-9849-4bd4-8976-38cc771e868e-kube-api-access-rqfl8\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095831 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bkkl\" (UniqueName: \"kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095857 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095883 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-dir\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095911 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-policies\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095942 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/000fdf2c-16b4-4271-a936-3795037adc6e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.095976 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096005 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096229 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6ww7\" (UniqueName: \"kubernetes.io/projected/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-kube-api-access-m6ww7\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096264 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096291 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/000fdf2c-16b4-4271-a936-3795037adc6e-serving-cert\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096319 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096346 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zczkp\" (UniqueName: \"kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096381 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-machine-approver-tls\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096408 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-service-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096436 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096463 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-client\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096490 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096517 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096542 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-serving-cert\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096610 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096646 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096694 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096723 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096767 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096800 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4cb\" (UniqueName: \"kubernetes.io/projected/adba3972-5deb-41ab-adcf-53795715f713-kube-api-access-5r4cb\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096828 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096855 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-image-import-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096905 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7277\" (UniqueName: \"kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096932 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096963 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.096990 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097020 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-serving-cert\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097065 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm5wt\" (UniqueName: \"kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097090 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097127 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmbwt\" (UniqueName: \"kubernetes.io/projected/b1bcc8c9-7f94-4733-b691-53f0e30df85c-kube-api-access-wmbwt\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097150 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097178 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-encryption-config\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097203 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxxx\" (UniqueName: \"kubernetes.io/projected/a71a5c20-c130-4d0d-9bef-f4c12204ae67-kube-api-access-2kxxx\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097232 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097258 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-encryption-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097289 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097315 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d88b\" (UniqueName: \"kubernetes.io/projected/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-kube-api-access-5d88b\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097340 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097364 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-config\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097412 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097442 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qrsv\" (UniqueName: \"kubernetes.io/projected/000fdf2c-16b4-4271-a936-3795037adc6e-kube-api-access-2qrsv\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097465 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097488 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-node-pullsecrets\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097512 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097542 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097585 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097616 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.097645 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-serving-cert\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.099587 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.099808 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.100510 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-294fg"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.101148 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.101286 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.101462 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.102068 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.104653 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.105137 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.105494 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.105965 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.106450 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.106678 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.110537 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.110955 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vfvmf"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.126326 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.130627 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.139945 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.140069 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.141251 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.143150 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.143901 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.145964 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.146963 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2gvn"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.147468 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.147834 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.148034 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.151460 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.152528 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.153245 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558890-7spp4"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.154306 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lvzfq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.154441 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.154882 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.155368 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.155513 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4c7wr"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.161234 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zw4vw"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.174249 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.178721 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xzvqd"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.180364 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.192219 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.192283 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hjq5h"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.193381 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.196516 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.196657 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199368 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199439 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-image-import-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199493 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7277\" (UniqueName: \"kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199544 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5wt\" (UniqueName: \"kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199600 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmbwt\" (UniqueName: \"kubernetes.io/projected/b1bcc8c9-7f94-4733-b691-53f0e30df85c-kube-api-access-wmbwt\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199626 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1901c2b3-404c-4363-a77d-724f71aedb02-serving-cert\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199665 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89wtl\" (UniqueName: \"kubernetes.io/projected/8e903079-fff6-462b-a805-3dac5d18a169-kube-api-access-89wtl\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199694 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199734 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-encryption-config\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199765 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxxx\" (UniqueName: \"kubernetes.io/projected/a71a5c20-c130-4d0d-9bef-f4c12204ae67-kube-api-access-2kxxx\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199790 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fbab285-e809-4147-846b-bd2252bf4f18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199834 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/164464a6-d9c6-475e-a7ae-707365c5ec51-serving-cert\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199859 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b9027a-9381-417c-b500-16822f802828-trusted-ca\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199904 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-node-pullsecrets\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199929 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199947 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09c6e2b5-3909-400a-a052-5e540226011b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.199986 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200007 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200026 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200060 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-serving-cert\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200080 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfz4\" (UniqueName: \"kubernetes.io/projected/116d02ff-db52-4e72-92a5-9b2de33db24a-kube-api-access-8dfz4\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200098 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-client\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200134 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2drp\" (UniqueName: \"kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200153 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd65149-6fb3-492b-9719-e5e22e51eedd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200222 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200248 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200281 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-metrics-certs\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200299 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116d02ff-db52-4e72-92a5-9b2de33db24a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200317 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200333 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200378 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drcws\" (UniqueName: \"kubernetes.io/projected/b0c9069d-5138-4fd9-9c2f-78b5224287e8-kube-api-access-drcws\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200413 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44h8w\" (UniqueName: \"kubernetes.io/projected/2ec532f5-c27d-4a38-900c-fbbc5676a5cd-kube-api-access-44h8w\") pod \"downloads-7954f5f757-xzvqd\" (UID: \"2ec532f5-c27d-4a38-900c-fbbc5676a5cd\") " pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200461 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200482 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit-dir\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200497 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqhsv\" (UniqueName: \"kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200545 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200590 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200611 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200629 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-dir\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200669 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200687 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx6g8\" (UniqueName: \"kubernetes.io/projected/381e7ca1-1798-42ef-8df1-bc0d15f88e23-kube-api-access-hx6g8\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200704 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4wkq\" (UniqueName: \"kubernetes.io/projected/43636426-c477-4b47-b209-87df8d3414bc-kube-api-access-s4wkq\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200721 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6ww7\" (UniqueName: \"kubernetes.io/projected/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-kube-api-access-m6ww7\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200753 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5569250f-d178-41af-ac06-64aa34373900-tmpfs\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200770 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pvcr\" (UniqueName: \"kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200786 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlnz6\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-kube-api-access-mlnz6\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200805 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5rpr\" (UniqueName: \"kubernetes.io/projected/5569250f-d178-41af-ac06-64aa34373900-kube-api-access-z5rpr\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200852 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200903 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200922 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s8dn\" (UniqueName: \"kubernetes.io/projected/1901c2b3-404c-4363-a77d-724f71aedb02-kube-api-access-9s8dn\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200941 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200959 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zczkp\" (UniqueName: \"kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.200992 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-service-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201012 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-client\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201041 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s66wm\" (UniqueName: \"kubernetes.io/projected/164464a6-d9c6-475e-a7ae-707365c5ec51-kube-api-access-s66wm\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201075 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201091 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a2b9027a-9381-417c-b500-16822f802828-metrics-tls\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201159 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201204 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201231 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201423 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201244 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201491 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201509 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201549 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201591 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201610 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4cb\" (UniqueName: \"kubernetes.io/projected/adba3972-5deb-41ab-adcf-53795715f713-kube-api-access-5r4cb\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201628 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.201694 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sdg6"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.202090 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit-dir\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203238 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203286 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203316 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-dir\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203383 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203450 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-image-import-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.203860 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.206355 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.206970 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-service-ca\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.206968 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.207299 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-node-pullsecrets\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.209237 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.209871 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.210980 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211788 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28681c01-d949-429f-ad5c-6084f6d30e81-config\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211840 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-serving-cert\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211888 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211907 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211926 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211961 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.211987 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktqxc\" (UniqueName: \"kubernetes.io/projected/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-kube-api-access-ktqxc\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212006 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116d02ff-db52-4e72-92a5-9b2de33db24a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212054 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-encryption-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212077 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212114 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212138 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d88b\" (UniqueName: \"kubernetes.io/projected/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-kube-api-access-5d88b\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212163 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28681c01-d949-429f-ad5c-6084f6d30e81-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212218 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212253 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-config\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212306 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tntnf\" (UniqueName: \"kubernetes.io/projected/09c6e2b5-3909-400a-a052-5e540226011b-kube-api-access-tntnf\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212385 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212421 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212460 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qrsv\" (UniqueName: \"kubernetes.io/projected/000fdf2c-16b4-4271-a936-3795037adc6e-kube-api-access-2qrsv\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212492 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212535 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-config\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212580 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-service-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212601 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nljdm\" (UniqueName: \"kubernetes.io/projected/631708f5-e5a1-4d6f-8858-60ea4de43cd9-kube-api-access-nljdm\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212622 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxr7j\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-kube-api-access-nxr7j\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212670 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-default-certificate\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212693 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjcpt\" (UniqueName: \"kubernetes.io/projected/0f7ab638-daa2-4d39-80e7-6a8061f62357-kube-api-access-qjcpt\") pod \"migrator-59844c95c7-s9k7b\" (UID: \"0f7ab638-daa2-4d39-80e7-6a8061f62357\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212712 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212756 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212784 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212830 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dd65149-6fb3-492b-9719-e5e22e51eedd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212850 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1901c2b3-404c-4363-a77d-724f71aedb02-config\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212869 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212904 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-auth-proxy-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212927 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09c6e2b5-3909-400a-a052-5e540226011b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212947 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.212986 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213012 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-client\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213030 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfl8\" (UniqueName: \"kubernetes.io/projected/6ef26501-9849-4bd4-8976-38cc771e868e-kube-api-access-rqfl8\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213064 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bkkl\" (UniqueName: \"kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213082 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213098 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-policies\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213132 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213151 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213166 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/000fdf2c-16b4-4271-a936-3795037adc6e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213184 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e903079-fff6-462b-a805-3dac5d18a169-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213219 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213236 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dd65149-6fb3-492b-9719-e5e22e51eedd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213253 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/000fdf2c-16b4-4271-a936-3795037adc6e-serving-cert\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213294 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213317 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-stats-auth\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213334 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213373 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-machine-approver-tls\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213394 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213413 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8fbab285-e809-4147-846b-bd2252bf4f18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213450 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213471 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdf7w\" (UniqueName: \"kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w\") pod \"auto-csr-approver-29558890-7spp4\" (UID: \"4585669d-fc84-4c80-9cc3-0185b7fb9ce0\") " pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213488 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213507 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213551 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213600 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-serving-cert\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213621 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631708f5-e5a1-4d6f-8858-60ea4de43cd9-service-ca-bundle\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213672 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213691 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqvp4\" (UniqueName: \"kubernetes.io/projected/997eb245-8ce2-4eba-a167-444b69f138c2-kube-api-access-bqvp4\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213711 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213754 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e903079-fff6-462b-a805-3dac5d18a169-proxy-tls\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213774 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-bound-sa-token\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.215336 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-serving-cert\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.215905 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a71a5c20-c130-4d0d-9bef-f4c12204ae67-config\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.216945 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.217466 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-client\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.217852 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/000fdf2c-16b4-4271-a936-3795037adc6e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.217877 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b1bcc8c9-7f94-4733-b691-53f0e30df85c-audit-policies\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.222134 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/000fdf2c-16b4-4271-a936-3795037adc6e-serving-cert\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.222904 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-etcd-serving-ca\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.223681 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-audit\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.227022 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.213669 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.227259 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-auth-proxy-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.230722 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-serving-cert\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.231491 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a71a5c20-c130-4d0d-9bef-f4c12204ae67-etcd-client\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.233907 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-config\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.239268 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-serving-cert\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.239627 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-machine-approver-tls\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.249103 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.249171 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-etcd-client\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.249253 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.250475 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29558880-xw69x"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.251694 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-294fg"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.253807 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-encryption-config\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.255374 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.257894 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k9d4t"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.259639 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.260071 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b1bcc8c9-7f94-4733-b691-53f0e30df85c-encryption-config\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.260178 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.261538 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-k6lv8"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.262618 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-zqb9z"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.262692 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.263230 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.264650 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.264932 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.269106 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j2krj"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.269148 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lns4z"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.269163 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.271609 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.273148 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.273173 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.274367 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-phl6h"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.274878 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.277678 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.277714 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbnkg"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.277731 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.281592 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.281620 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.281665 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zqb9z"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.283040 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.286116 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2gvn"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.286142 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k6lv8"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.286157 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.289643 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.289767 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vfvmf"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.290243 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558890-7spp4"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.309769 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.310468 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sdg6"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.312276 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.317512 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.317467 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqvp4\" (UniqueName: \"kubernetes.io/projected/997eb245-8ce2-4eba-a167-444b69f138c2-kube-api-access-bqvp4\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320177 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631708f5-e5a1-4d6f-8858-60ea4de43cd9-service-ca-bundle\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320264 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e903079-fff6-462b-a805-3dac5d18a169-proxy-tls\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320304 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-bound-sa-token\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320420 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1901c2b3-404c-4363-a77d-724f71aedb02-serving-cert\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320503 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89wtl\" (UniqueName: \"kubernetes.io/projected/8e903079-fff6-462b-a805-3dac5d18a169-kube-api-access-89wtl\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320641 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fbab285-e809-4147-846b-bd2252bf4f18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320702 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/164464a6-d9c6-475e-a7ae-707365c5ec51-serving-cert\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320736 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b9027a-9381-417c-b500-16822f802828-trusted-ca\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.320772 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09c6e2b5-3909-400a-a052-5e540226011b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.321206 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfz4\" (UniqueName: \"kubernetes.io/projected/116d02ff-db52-4e72-92a5-9b2de33db24a-kube-api-access-8dfz4\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.321719 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd65149-6fb3-492b-9719-e5e22e51eedd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.321825 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2drp\" (UniqueName: \"kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.321884 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-metrics-certs\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.322113 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116d02ff-db52-4e72-92a5-9b2de33db24a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.322442 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.322504 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323020 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631708f5-e5a1-4d6f-8858-60ea4de43cd9-service-ca-bundle\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323094 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx6g8\" (UniqueName: \"kubernetes.io/projected/381e7ca1-1798-42ef-8df1-bc0d15f88e23-kube-api-access-hx6g8\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323143 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5569250f-d178-41af-ac06-64aa34373900-tmpfs\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323210 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pvcr\" (UniqueName: \"kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323465 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4wkq\" (UniqueName: \"kubernetes.io/projected/43636426-c477-4b47-b209-87df8d3414bc-kube-api-access-s4wkq\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323515 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5rpr\" (UniqueName: \"kubernetes.io/projected/5569250f-d178-41af-ac06-64aa34373900-kube-api-access-z5rpr\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323588 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlnz6\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-kube-api-access-mlnz6\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323620 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323647 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s8dn\" (UniqueName: \"kubernetes.io/projected/1901c2b3-404c-4363-a77d-724f71aedb02-kube-api-access-9s8dn\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323725 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s66wm\" (UniqueName: \"kubernetes.io/projected/164464a6-d9c6-475e-a7ae-707365c5ec51-kube-api-access-s66wm\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323804 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323905 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a2b9027a-9381-417c-b500-16822f802828-metrics-tls\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323952 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.323991 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324037 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324118 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28681c01-d949-429f-ad5c-6084f6d30e81-config\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324225 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324261 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktqxc\" (UniqueName: \"kubernetes.io/projected/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-kube-api-access-ktqxc\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324305 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116d02ff-db52-4e72-92a5-9b2de33db24a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324357 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28681c01-d949-429f-ad5c-6084f6d30e81-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324386 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324546 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tntnf\" (UniqueName: \"kubernetes.io/projected/09c6e2b5-3909-400a-a052-5e540226011b-kube-api-access-tntnf\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324627 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-config\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324646 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/116d02ff-db52-4e72-92a5-9b2de33db24a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324754 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-service-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324804 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nljdm\" (UniqueName: \"kubernetes.io/projected/631708f5-e5a1-4d6f-8858-60ea4de43cd9-kube-api-access-nljdm\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324840 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/5569250f-d178-41af-ac06-64aa34373900-tmpfs\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324872 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxr7j\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-kube-api-access-nxr7j\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325297 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-default-certificate\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325365 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324938 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325401 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjcpt\" (UniqueName: \"kubernetes.io/projected/0f7ab638-daa2-4d39-80e7-6a8061f62357-kube-api-access-qjcpt\") pod \"migrator-59844c95c7-s9k7b\" (UID: \"0f7ab638-daa2-4d39-80e7-6a8061f62357\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325434 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325531 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09c6e2b5-3909-400a-a052-5e540226011b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325642 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dd65149-6fb3-492b-9719-e5e22e51eedd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325742 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1901c2b3-404c-4363-a77d-724f71aedb02-config\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325782 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.325933 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326013 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e903079-fff6-462b-a805-3dac5d18a169-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.324944 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326064 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dd65149-6fb3-492b-9719-e5e22e51eedd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326096 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326157 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-stats-auth\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326314 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326360 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326392 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8fbab285-e809-4147-846b-bd2252bf4f18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326406 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-metrics-certs\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326421 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdf7w\" (UniqueName: \"kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w\") pod \"auto-csr-approver-29558890-7spp4\" (UID: \"4585669d-fc84-4c80-9cc3-0185b7fb9ce0\") " pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326515 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.326579 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.327700 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-config\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.327802 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e903079-fff6-462b-a805-3dac5d18a169-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.328199 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p"] Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.333178 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-stats-auth\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.333663 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/116d02ff-db52-4e72-92a5-9b2de33db24a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.335466 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/631708f5-e5a1-4d6f-8858-60ea4de43cd9-default-certificate\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.338788 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.348472 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/164464a6-d9c6-475e-a7ae-707365c5ec51-serving-cert\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.365261 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.366541 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.378545 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.388009 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/164464a6-d9c6-475e-a7ae-707365c5ec51-service-ca-bundle\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.399117 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.419424 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.439063 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.459038 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.478929 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.499658 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.504754 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09c6e2b5-3909-400a-a052-5e540226011b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.519630 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.527128 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09c6e2b5-3909-400a-a052-5e540226011b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.539502 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.558804 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.585411 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.592800 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fbab285-e809-4147-846b-bd2252bf4f18-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.601277 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.619711 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.639410 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.659755 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.671181 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/8fbab285-e809-4147-846b-bd2252bf4f18-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.679947 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.709807 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.720672 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.743533 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.779045 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.785816 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8e903079-fff6-462b-a805-3dac5d18a169-proxy-tls\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.798926 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.819524 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.831110 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.840242 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.859690 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.868270 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.881811 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.886700 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.911112 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.917288 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.919498 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.939995 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.959462 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.979720 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 15 00:10:56 crc kubenswrapper[4861]: I0315 00:10:56.999009 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.012054 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a2b9027a-9381-417c-b500-16822f802828-metrics-tls\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.029509 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.033715 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2b9027a-9381-417c-b500-16822f802828-trusted-ca\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.040770 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.060197 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.079604 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.099933 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.107332 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1901c2b3-404c-4363-a77d-724f71aedb02-serving-cert\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.119240 4861 request.go:700] Waited for 1.016884558s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.122648 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.128237 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1901c2b3-404c-4363-a77d-724f71aedb02-config\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.139610 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.160390 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.179736 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.190986 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5dd65149-6fb3-492b-9719-e5e22e51eedd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.199386 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202172 4861 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202283 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.702245846 +0000 UTC m=+251.014443609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202287 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202368 4861 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202413 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.70237558 +0000 UTC m=+251.014573123 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.202460 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.702430231 +0000 UTC m=+251.014627804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.203537 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dd65149-6fb3-492b-9719-e5e22e51eedd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.204741 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.204783 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.204838 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.704812114 +0000 UTC m=+251.017009848 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.204908 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.704853346 +0000 UTC m=+251.017051079 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208328 4861 configmap.go:193] Couldn't get configMap openshift-console/console-config: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208414 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208428 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.708397081 +0000 UTC m=+251.020594864 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-config" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208476 4861 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208478 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.708462333 +0000 UTC m=+251.020659876 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208617 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.708603517 +0000 UTC m=+251.020801060 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.208329 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-router-certs: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.209260 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.709214273 +0000 UTC m=+251.021411836 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-router-certs" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.209402 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.209679 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.709658375 +0000 UTC m=+251.021855948 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.209444 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.210086 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.710067226 +0000 UTC m=+251.022264789 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.210162 4861 secret.go:188] Couldn't get secret openshift-apiserver-operator/openshift-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.210399 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert podName:adba3972-5deb-41ab-adcf-53795715f713 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.710383244 +0000 UTC m=+251.022580817 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert") pod "openshift-apiserver-operator-796bbdcf4f-kmj2d" (UID: "adba3972-5deb-41ab-adcf-53795715f713") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.213098 4861 configmap.go:193] Couldn't get configMap openshift-console/service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.213203 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.713176739 +0000 UTC m=+251.025374462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215121 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215199 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.715180453 +0000 UTC m=+251.027378026 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215204 4861 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215487 4861 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215547 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.715534113 +0000 UTC m=+251.027731666 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.215593 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.715578944 +0000 UTC m=+251.027776487 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.216805 4861 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.216894 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls podName:432faa21-42d4-454a-a36f-dcf427e79e96 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.716869609 +0000 UTC m=+251.029067182 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-7jkmx" (UID: "432faa21-42d4-454a-a36f-dcf427e79e96") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.217058 4861 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.217646 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config podName:adba3972-5deb-41ab-adcf-53795715f713 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.717622708 +0000 UTC m=+251.029820271 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config") pod "openshift-apiserver-operator-796bbdcf4f-kmj2d" (UID: "adba3972-5deb-41ab-adcf-53795715f713") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.217286 4861 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.217425 4861 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.217463 4861 secret.go:188] Couldn't get secret openshift-dns-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.219009 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223131 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.723051844 +0000 UTC m=+251.035249577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223256 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.723231889 +0000 UTC m=+251.035429452 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223351 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls podName:3589ff40-d99e-4598-8f93-b8721775f117 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.723335363 +0000 UTC m=+251.035532936 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls") pod "dns-operator-744455d44c-lns4z" (UID: "3589ff40-d99e-4598-8f93-b8721775f117") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.219059 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223525 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.723465126 +0000 UTC m=+251.035662939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223625 4861 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.223821 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.723800135 +0000 UTC m=+251.035997698 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.228750 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.228833 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.728813779 +0000 UTC m=+251.041011352 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.228872 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.228913 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.728900582 +0000 UTC m=+251.041098145 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.240398 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.248594 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28681c01-d949-429f-ad5c-6084f6d30e81-config\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.259199 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.278775 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.298994 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.318415 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.322942 4861 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.323070 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume podName:59b39d47-b227-436a-9437-688c54bd4e05 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.823040249 +0000 UTC m=+251.135237802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume") pod "collect-profiles-29558880-stl52" (UID: "59b39d47-b227-436a-9437-688c54bd4e05") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.325274 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.325347 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume podName:59b39d47-b227-436a-9437-688c54bd4e05 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.82532982 +0000 UTC m=+251.137527373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume") pod "collect-profiles-29558880-stl52" (UID: "59b39d47-b227-436a-9437-688c54bd4e05") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.325777 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.325870 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert podName:5569250f-d178-41af-ac06-64aa34373900 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.825847454 +0000 UTC m=+251.138045027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert") pod "packageserver-d55dfcdfc-rmkcs" (UID: "5569250f-d178-41af-ac06-64aa34373900") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.325937 4861 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.326114 4861 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.326193 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle podName:a861f914-a0a6-4b7e-98b7-94cbc3ea942d nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.826174553 +0000 UTC m=+251.138372126 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle") pod "service-ca-9c57cc56f-z2gvn" (UID: "a861f914-a0a6-4b7e-98b7-94cbc3ea942d") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.326265 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert podName:28681c01-d949-429f-ad5c-6084f6d30e81 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.826232024 +0000 UTC m=+251.138429567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert") pod "kube-apiserver-operator-766d6c64bb-dcj6k" (UID: "28681c01-d949-429f-ad5c-6084f6d30e81") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327661 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327723 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327763 4861 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327808 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327737 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert podName:381e7ca1-1798-42ef-8df1-bc0d15f88e23 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.827715694 +0000 UTC m=+251.139913437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert") pod "olm-operator-6b444d44fb-bk6w5" (UID: "381e7ca1-1798-42ef-8df1-bc0d15f88e23") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327853 4861 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327873 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert podName:43636426-c477-4b47-b209-87df8d3414bc nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.827852288 +0000 UTC m=+251.140050041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert") pod "catalog-operator-68c6474976-7hrpb" (UID: "43636426-c477-4b47-b209-87df8d3414bc") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327902 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key podName:a861f914-a0a6-4b7e-98b7-94cbc3ea942d nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.827887999 +0000 UTC m=+251.140085542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key") pod "service-ca-9c57cc56f-z2gvn" (UID: "a861f914-a0a6-4b7e-98b7-94cbc3ea942d") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327931 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs podName:997eb245-8ce2-4eba-a167-444b69f138c2 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.827918059 +0000 UTC m=+251.140115612 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs") pod "multus-admission-controller-857f4d67dd-vfvmf" (UID: "997eb245-8ce2-4eba-a167-444b69f138c2") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.327961 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert podName:43636426-c477-4b47-b209-87df8d3414bc nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.82794505 +0000 UTC m=+251.140142603 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert") pod "catalog-operator-68c6474976-7hrpb" (UID: "43636426-c477-4b47-b209-87df8d3414bc") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.328469 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.328518 4861 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.328529 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert podName:381e7ca1-1798-42ef-8df1-bc0d15f88e23 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.828513546 +0000 UTC m=+251.140711289 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert") pod "olm-operator-6b444d44fb-bk6w5" (UID: "381e7ca1-1798-42ef-8df1-bc0d15f88e23") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: E0315 00:10:57.328677 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert podName:5569250f-d178-41af-ac06-64aa34373900 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:57.82865767 +0000 UTC m=+251.140855243 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert") pod "packageserver-d55dfcdfc-rmkcs" (UID: "5569250f-d178-41af-ac06-64aa34373900") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.338939 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.358856 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.378615 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.398082 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.418824 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.438889 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.459717 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.478680 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.521343 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.521921 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.539337 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.559540 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.579874 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.599071 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.619314 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.638777 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.659030 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.679759 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.698847 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.720423 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.740162 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.759699 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761238 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761306 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761352 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761380 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761430 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761456 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761520 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761673 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761711 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761752 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761790 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761820 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761880 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761906 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761949 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.761992 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762036 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762082 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762156 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762191 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762231 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762280 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762376 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762422 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.762449 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.779744 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.799257 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.818322 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.839144 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.859448 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.864061 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.864259 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.864452 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.864524 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.864712 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865018 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865087 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865201 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865674 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865811 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865844 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-cabundle\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865859 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.865960 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.866007 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.870866 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.871977 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.872499 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-signing-key\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.873395 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/381e7ca1-1798-42ef-8df1-bc0d15f88e23-srv-cert\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.874281 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-profile-collector-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.875134 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-apiservice-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.876114 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/43636426-c477-4b47-b209-87df8d3414bc-srv-cert\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.876125 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5569250f-d178-41af-ac06-64aa34373900-webhook-cert\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.876282 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/997eb245-8ce2-4eba-a167-444b69f138c2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.876912 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28681c01-d949-429f-ad5c-6084f6d30e81-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.899423 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.919418 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.938810 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 15 00:10:57 crc kubenswrapper[4861]: I0315 00:10:57.989161 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drcws\" (UniqueName: \"kubernetes.io/projected/b0c9069d-5138-4fd9-9c2f-78b5224287e8-kube-api-access-drcws\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.008414 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44h8w\" (UniqueName: \"kubernetes.io/projected/2ec532f5-c27d-4a38-900c-fbbc5676a5cd-kube-api-access-44h8w\") pod \"downloads-7954f5f757-xzvqd\" (UID: \"2ec532f5-c27d-4a38-900c-fbbc5676a5cd\") " pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.020380 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.060067 4861 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.079231 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.137482 4861 request.go:700] Waited for 1.933453002s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.149237 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmbwt\" (UniqueName: \"kubernetes.io/projected/b1bcc8c9-7f94-4733-b691-53f0e30df85c-kube-api-access-wmbwt\") pod \"apiserver-7bbb656c7d-p295d\" (UID: \"b1bcc8c9-7f94-4733-b691-53f0e30df85c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.155075 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6ww7\" (UniqueName: \"kubernetes.io/projected/009b6cf3-3e73-4b27-b249-1c72de7bd3dd-kube-api-access-m6ww7\") pod \"apiserver-76f77b778f-zw4vw\" (UID: \"009b6cf3-3e73-4b27-b249-1c72de7bd3dd\") " pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.155422 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.182340 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zczkp\" (UniqueName: \"kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp\") pod \"route-controller-manager-6576b87f9c-5fg69\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.207699 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxxx\" (UniqueName: \"kubernetes.io/projected/a71a5c20-c130-4d0d-9bef-f4c12204ae67-kube-api-access-2kxxx\") pod \"etcd-operator-b45778765-lvzfq\" (UID: \"a71a5c20-c130-4d0d-9bef-f4c12204ae67\") " pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.251022 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.260626 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d88b\" (UniqueName: \"kubernetes.io/projected/0a957ae9-6295-4228-9dfc-6f50cdfef5e2-kube-api-access-5d88b\") pod \"machine-approver-56656f9798-8pv8j\" (UID: \"0a957ae9-6295-4228-9dfc-6f50cdfef5e2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.263929 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.278513 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.298797 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qrsv\" (UniqueName: \"kubernetes.io/projected/000fdf2c-16b4-4271-a936-3795037adc6e-kube-api-access-2qrsv\") pod \"openshift-config-operator-7777fb866f-dh2q9\" (UID: \"000fdf2c-16b4-4271-a936-3795037adc6e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.300240 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.305162 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bkkl\" (UniqueName: \"kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl\") pod \"image-pruner-29558880-xw69x\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.319859 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.342364 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.359920 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.373763 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.379057 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.400784 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.420018 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.460706 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.464862 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.475932 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89wtl\" (UniqueName: \"kubernetes.io/projected/8e903079-fff6-462b-a805-3dac5d18a169-kube-api-access-89wtl\") pod \"machine-config-controller-84d6567774-8chcv\" (UID: \"8e903079-fff6-462b-a805-3dac5d18a169\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.481476 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqvp4\" (UniqueName: \"kubernetes.io/projected/997eb245-8ce2-4eba-a167-444b69f138c2-kube-api-access-bqvp4\") pod \"multus-admission-controller-857f4d67dd-vfvmf\" (UID: \"997eb245-8ce2-4eba-a167-444b69f138c2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.495381 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-bound-sa-token\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.514097 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfz4\" (UniqueName: \"kubernetes.io/projected/116d02ff-db52-4e72-92a5-9b2de33db24a-kube-api-access-8dfz4\") pod \"kube-storage-version-migrator-operator-b67b599dd-dbghq\" (UID: \"116d02ff-db52-4e72-92a5-9b2de33db24a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.538643 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.542804 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.543064 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.555932 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2drp\" (UniqueName: \"kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp\") pod \"controller-manager-879f6c89f-8pnz7\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.578808 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" event={"ID":"b1bcc8c9-7f94-4733-b691-53f0e30df85c","Type":"ContainerStarted","Data":"5549731c25679a3f3de23c7d304927d86dd404e9ea8e44735a0fc18cd4c3b74a"} Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.580104 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx6g8\" (UniqueName: \"kubernetes.io/projected/381e7ca1-1798-42ef-8df1-bc0d15f88e23-kube-api-access-hx6g8\") pod \"olm-operator-6b444d44fb-bk6w5\" (UID: \"381e7ca1-1798-42ef-8df1-bc0d15f88e23\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.588786 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.594056 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pvcr\" (UniqueName: \"kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr\") pod \"collect-profiles-29558880-stl52\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.611705 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.615987 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4wkq\" (UniqueName: \"kubernetes.io/projected/43636426-c477-4b47-b209-87df8d3414bc-kube-api-access-s4wkq\") pod \"catalog-operator-68c6474976-7hrpb\" (UID: \"43636426-c477-4b47-b209-87df8d3414bc\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.634716 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.636360 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5rpr\" (UniqueName: \"kubernetes.io/projected/5569250f-d178-41af-ac06-64aa34373900-kube-api-access-z5rpr\") pod \"packageserver-d55dfcdfc-rmkcs\" (UID: \"5569250f-d178-41af-ac06-64aa34373900\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.637018 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-zw4vw"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.655333 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlnz6\" (UniqueName: \"kubernetes.io/projected/8fbab285-e809-4147-846b-bd2252bf4f18-kube-api-access-mlnz6\") pod \"cluster-image-registry-operator-dc59b4c8b-p96db\" (UID: \"8fbab285-e809-4147-846b-bd2252bf4f18\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.661486 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod009b6cf3_3e73_4b27_b249_1c72de7bd3dd.slice/crio-52ff1225d2585860d78a928c5d2708f891832369fa028120510cd2340994352a WatchSource:0}: Error finding container 52ff1225d2585860d78a928c5d2708f891832369fa028120510cd2340994352a: Status 404 returned error can't find the container with id 52ff1225d2585860d78a928c5d2708f891832369fa028120510cd2340994352a Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.661821 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.671965 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.676542 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29558880-xw69x"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.688490 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s8dn\" (UniqueName: \"kubernetes.io/projected/1901c2b3-404c-4363-a77d-724f71aedb02-kube-api-access-9s8dn\") pod \"service-ca-operator-777779d784-hrl6d\" (UID: \"1901c2b3-404c-4363-a77d-724f71aedb02\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.695916 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s66wm\" (UniqueName: \"kubernetes.io/projected/164464a6-d9c6-475e-a7ae-707365c5ec51-kube-api-access-s66wm\") pod \"authentication-operator-69f744f599-phl6h\" (UID: \"164464a6-d9c6-475e-a7ae-707365c5ec51\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.701330 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xzvqd"] Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.716413 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod981ba86e_1a94_4acc_ae13_25da019d91fd.slice/crio-d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da WatchSource:0}: Error finding container d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da: Status 404 returned error can't find the container with id d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.717321 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28681c01-d949-429f-ad5c-6084f6d30e81-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dcj6k\" (UID: \"28681c01-d949-429f-ad5c-6084f6d30e81\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.738814 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktqxc\" (UniqueName: \"kubernetes.io/projected/a861f914-a0a6-4b7e-98b7-94cbc3ea942d-kube-api-access-ktqxc\") pod \"service-ca-9c57cc56f-z2gvn\" (UID: \"a861f914-a0a6-4b7e-98b7-94cbc3ea942d\") " pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.759679 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nljdm\" (UniqueName: \"kubernetes.io/projected/631708f5-e5a1-4d6f-8858-60ea4de43cd9-kube-api-access-nljdm\") pod \"router-default-5444994796-lgbxx\" (UID: \"631708f5-e5a1-4d6f-8858-60ea4de43cd9\") " pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762247 4861 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762316 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762292935 +0000 UTC m=+253.074490458 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.762397 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-lvzfq"] Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762513 4861 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762537 4861 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762568 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762603 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762541 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762534722 +0000 UTC m=+253.074732255 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762690 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762657175 +0000 UTC m=+253.074854888 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762690 4861 configmap.go:193] Couldn't get configMap openshift-console/console-config: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762717 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762707306 +0000 UTC m=+253.074905059 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762738 4861 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762744 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762733267 +0000 UTC m=+253.074931020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762638 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762764 4861 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762766 4861 secret.go:188] Couldn't get secret openshift-apiserver-operator/openshift-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762791 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762781858 +0000 UTC m=+253.074979611 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762805 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-router-certs: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762807 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762798959 +0000 UTC m=+253.074996702 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762782 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762845 4861 configmap.go:193] Couldn't get configMap openshift-console/service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762879 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762894 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762707 4861 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762925 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762949 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762826 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762820849 +0000 UTC m=+253.075018382 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-router-certs" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762969 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762963683 +0000 UTC m=+253.075161216 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.762985 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert podName:adba3972-5deb-41ab-adcf-53795715f713 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762978733 +0000 UTC m=+253.075176496 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert") pod "openshift-apiserver-operator-796bbdcf4f-kmj2d" (UID: "adba3972-5deb-41ab-adcf-53795715f713") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763001 4861 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763011 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.762993904 +0000 UTC m=+253.075191437 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763028 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763019314 +0000 UTC m=+253.075216847 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763043 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763038365 +0000 UTC m=+253.075235888 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763067 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763061175 +0000 UTC m=+253.075258708 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763081 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763074676 +0000 UTC m=+253.075272439 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763099 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763091876 +0000 UTC m=+253.075289409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763113 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763106097 +0000 UTC m=+253.075303620 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763136 4861 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763142 4861 secret.go:188] Couldn't get secret openshift-dns-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763156 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images podName:6ef26501-9849-4bd4-8976-38cc771e868e nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763151438 +0000 UTC m=+253.075348971 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images") pod "machine-api-operator-5694c8668f-fbnkg" (UID: "6ef26501-9849-4bd4-8976-38cc771e868e") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763175 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls podName:3589ff40-d99e-4598-8f93-b8721775f117 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763167288 +0000 UTC m=+253.075364821 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls") pod "dns-operator-744455d44c-lns4z" (UID: "3589ff40-d99e-4598-8f93-b8721775f117") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763203 4861 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763231 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config podName:adba3972-5deb-41ab-adcf-53795715f713 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.76322487 +0000 UTC m=+253.075422403 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config") pod "openshift-apiserver-operator-796bbdcf4f-kmj2d" (UID: "adba3972-5deb-41ab-adcf-53795715f713") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763256 4861 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763282 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls podName:432faa21-42d4-454a-a36f-dcf427e79e96 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763274001 +0000 UTC m=+253.075471534 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-7jkmx" (UID: "432faa21-42d4-454a-a36f-dcf427e79e96") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763315 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763288721 +0000 UTC m=+253.075486254 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.763338 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.763329772 +0000 UTC m=+253.075527305 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "console-config" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.764261 4861 configmap.go:193] Couldn't get configMap openshift-console/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.764332 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle podName:b0c9069d-5138-4fd9-9c2f-78b5224287e8 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.764308259 +0000 UTC m=+253.076505792 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle") pod "console-f9d7485db-4c7wr" (UID: "b0c9069d-5138-4fd9-9c2f-78b5224287e8") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.764357 4861 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: E0315 00:10:58.764392 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data podName:1e8992d1-2a3f-4a3b-b857-821cfb8bed31 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.764384911 +0000 UTC m=+253.076582444 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-j2krj" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31") : failed to sync secret cache: timed out waiting for the condition Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.766330 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.767150 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.774541 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.776036 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tntnf\" (UniqueName: \"kubernetes.io/projected/09c6e2b5-3909-400a-a052-5e540226011b-kube-api-access-tntnf\") pod \"openshift-controller-manager-operator-756b6f6bc6-4h9nq\" (UID: \"09c6e2b5-3909-400a-a052-5e540226011b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.781277 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.798331 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjcpt\" (UniqueName: \"kubernetes.io/projected/0f7ab638-daa2-4d39-80e7-6a8061f62357-kube-api-access-qjcpt\") pod \"migrator-59844c95c7-s9k7b\" (UID: \"0f7ab638-daa2-4d39-80e7-6a8061f62357\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.803096 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.809057 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6174b71_a5cc_4a0e_902b_9a9091d15d83.slice/crio-9f7f90b1807b7285d7d41bcca4ae4a01b99a9a05065bce6155ff02068441cec9 WatchSource:0}: Error finding container 9f7f90b1807b7285d7d41bcca4ae4a01b99a9a05065bce6155ff02068441cec9: Status 404 returned error can't find the container with id 9f7f90b1807b7285d7d41bcca4ae4a01b99a9a05065bce6155ff02068441cec9 Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.809802 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.813788 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5dd65149-6fb3-492b-9719-e5e22e51eedd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-87hd7\" (UID: \"5dd65149-6fb3-492b-9719-e5e22e51eedd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.814050 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e903079_fff6_462b_a805_3dac5d18a169.slice/crio-892a8aa81430fb1f6698a52b3536ba7d46aab14d05ed87c42cc322603b66fd20 WatchSource:0}: Error finding container 892a8aa81430fb1f6698a52b3536ba7d46aab14d05ed87c42cc322603b66fd20: Status 404 returned error can't find the container with id 892a8aa81430fb1f6698a52b3536ba7d46aab14d05ed87c42cc322603b66fd20 Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.817163 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.829053 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda71a5c20_c130_4d0d_9bef_f4c12204ae67.slice/crio-39a412fec0453f5fd18184c222a325df1011970ebd936e19307e689a65fc69ed WatchSource:0}: Error finding container 39a412fec0453f5fd18184c222a325df1011970ebd936e19307e689a65fc69ed: Status 404 returned error can't find the container with id 39a412fec0453f5fd18184c222a325df1011970ebd936e19307e689a65fc69ed Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.833118 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.833400 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.841157 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdf7w\" (UniqueName: \"kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w\") pod \"auto-csr-approver-29558890-7spp4\" (UID: \"4585669d-fc84-4c80-9cc3-0185b7fb9ce0\") " pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.849518 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.856288 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxr7j\" (UniqueName: \"kubernetes.io/projected/a2b9027a-9381-417c-b500-16822f802828-kube-api-access-nxr7j\") pod \"ingress-operator-5b745b69d9-294fg\" (UID: \"a2b9027a-9381-417c-b500-16822f802828\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.857847 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.865952 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.873453 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.880789 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.887735 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.895324 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.903002 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vfvmf"] Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.904786 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.919835 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.943983 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.947251 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.954545 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.959495 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 15 00:10:58 crc kubenswrapper[4861]: W0315 00:10:58.968830 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997eb245_8ce2_4eba_a167_444b69f138c2.slice/crio-2ea957a27cf5173a40f5da82a6ef1fcf1ad64620e86c153e62764c6350a02f2c WatchSource:0}: Error finding container 2ea957a27cf5173a40f5da82a6ef1fcf1ad64620e86c153e62764c6350a02f2c: Status 404 returned error can't find the container with id 2ea957a27cf5173a40f5da82a6ef1fcf1ad64620e86c153e62764c6350a02f2c Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.981047 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 15 00:10:58 crc kubenswrapper[4861]: I0315 00:10:58.998382 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.018210 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.026972 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.040138 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.051842 4861 projected.go:288] Couldn't get configMap openshift-dns-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.060401 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.082197 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.095145 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.100049 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.115730 4861 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.115798 4861 projected.go:194] Error preparing data for projected volume kube-api-access-cm5wt for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.115899 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt podName:432faa21-42d4-454a-a36f-dcf427e79e96 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.615876157 +0000 UTC m=+252.928073690 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cm5wt" (UniqueName: "kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt") pod "cluster-samples-operator-665b6dd947-7jkmx" (UID: "432faa21-42d4-454a-a36f-dcf427e79e96") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.120577 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.141713 4861 request.go:700] Waited for 2.111937585s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/configmaps?fieldSelector=metadata.name%3Dv4-0-config-system-cliconfig&limit=500&resourceVersion=0 Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.144765 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.161007 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.185025 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4cb\" (UniqueName: \"kubernetes.io/projected/adba3972-5deb-41ab-adcf-53795715f713-kube-api-access-5r4cb\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.187165 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.190325 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.206894 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.208672 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.239942 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.254161 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.259370 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.270830 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.291192 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.299023 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.302210 4861 projected.go:194] Error preparing data for projected volume kube-api-access-p7277 for pod openshift-dns-operator/dns-operator-744455d44c-lns4z: failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.302299 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277 podName:3589ff40-d99e-4598-8f93-b8721775f117 nodeName:}" failed. No retries permitted until 2026-03-15 00:10:59.80227209 +0000 UTC m=+253.114469623 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-p7277" (UniqueName: "kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277") pod "dns-operator-744455d44c-lns4z" (UID: "3589ff40-d99e-4598-8f93-b8721775f117") : failed to sync configmap cache: timed out waiting for the condition Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.329690 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.350294 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.362274 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.370327 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfl8\" (UniqueName: \"kubernetes.io/projected/6ef26501-9849-4bd4-8976-38cc771e868e-kube-api-access-rqfl8\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.382664 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.382807 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.422251 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.429566 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.440495 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.467862 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.481926 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.499447 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.506990 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqhsv\" (UniqueName: \"kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.518782 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.540653 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.559009 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.579676 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.585974 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xzvqd" event={"ID":"2ec532f5-c27d-4a38-900c-fbbc5676a5cd","Type":"ContainerStarted","Data":"3580970b8121bb58c6007a0ba9e2c1e89a510417637c78c61b3f94151e4a05c3"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.586023 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xzvqd" event={"ID":"2ec532f5-c27d-4a38-900c-fbbc5676a5cd","Type":"ContainerStarted","Data":"eab463545c496ea48ef1452a05149477560af8138f4904ec5920f178e08cd70b"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.587029 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.588143 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" event={"ID":"997eb245-8ce2-4eba-a167-444b69f138c2","Type":"ContainerStarted","Data":"2ea957a27cf5173a40f5da82a6ef1fcf1ad64620e86c153e62764c6350a02f2c"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.589298 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" event={"ID":"0a957ae9-6295-4228-9dfc-6f50cdfef5e2","Type":"ContainerStarted","Data":"83321f7fb52e06c432f2c01ab8387063272839957c355cca321e9be1bc6c3491"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.589329 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" event={"ID":"0a957ae9-6295-4228-9dfc-6f50cdfef5e2","Type":"ContainerStarted","Data":"806e795efcedbf4531cfb0c0872c2a7ed84e0f8b8e20a72ea30ced159f2248eb"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.589958 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lgbxx" event={"ID":"631708f5-e5a1-4d6f-8858-60ea4de43cd9","Type":"ContainerStarted","Data":"114967a3feb79e8e5396ce4f0c540743ac22bbbd109655560604be69062b0adf"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.595619 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.595675 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.595954 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" event={"ID":"a71a5c20-c130-4d0d-9bef-f4c12204ae67","Type":"ContainerStarted","Data":"39a412fec0453f5fd18184c222a325df1011970ebd936e19307e689a65fc69ed"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.597987 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.599115 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" event={"ID":"c6174b71-a5cc-4a0e-902b-9a9091d15d83","Type":"ContainerStarted","Data":"9f7f90b1807b7285d7d41bcca4ae4a01b99a9a05065bce6155ff02068441cec9"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.602677 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" event={"ID":"43636426-c477-4b47-b209-87df8d3414bc","Type":"ContainerStarted","Data":"4074b04fbfd4c442c4614c480b34b84ea3db09f042484463e6d310795b9b9a79"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.604604 4861 generic.go:334] "Generic (PLEG): container finished" podID="b1bcc8c9-7f94-4733-b691-53f0e30df85c" containerID="c17637e5bca81a86f589ed1d244f905544c419371c9138018306e591d8ecb037" exitCode=0 Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.604664 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" event={"ID":"b1bcc8c9-7f94-4733-b691-53f0e30df85c","Type":"ContainerDied","Data":"c17637e5bca81a86f589ed1d244f905544c419371c9138018306e591d8ecb037"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.619706 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.624255 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" event={"ID":"5569250f-d178-41af-ac06-64aa34373900","Type":"ContainerStarted","Data":"36d21fb0d9da9e335b4daa762cc9f6058862ae15f4ec3df4fbdcafa554d5a8d1"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.628682 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm5wt\" (UniqueName: \"kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.630996 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" event={"ID":"8e903079-fff6-462b-a805-3dac5d18a169","Type":"ContainerStarted","Data":"892a8aa81430fb1f6698a52b3536ba7d46aab14d05ed87c42cc322603b66fd20"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.639216 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29558880-xw69x" event={"ID":"981ba86e-1a94-4acc-ae13-25da019d91fd","Type":"ContainerStarted","Data":"006be2d24ad8ed61eec2ba7fd726728425f2f04791bcf16bc7071441bd028fbc"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.639300 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29558880-xw69x" event={"ID":"981ba86e-1a94-4acc-ae13-25da019d91fd","Type":"ContainerStarted","Data":"d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.641331 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.642743 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" event={"ID":"116d02ff-db52-4e72-92a5-9b2de33db24a","Type":"ContainerStarted","Data":"af892b533570db62b408cb23291419f5991bc9c484916c19e31c352f54717971"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.651439 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" event={"ID":"59b39d47-b227-436a-9437-688c54bd4e05","Type":"ContainerStarted","Data":"7c815ad3ba32fd9296c648a3c28e56206d31191b6da8d82de0bbf92632ec0652"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.657693 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" event={"ID":"381e7ca1-1798-42ef-8df1-bc0d15f88e23","Type":"ContainerStarted","Data":"7a58f97d24b309da49930fd19afce51bf082858de1ee94537d6c35dda03f4a43"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.662229 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.666693 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" event={"ID":"0f7ab638-daa2-4d39-80e7-6a8061f62357","Type":"ContainerStarted","Data":"12a717069683ee5c1191672cfcf560f10b1f0c53311b915bca94f8d02e48e96b"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.678973 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" event={"ID":"000fdf2c-16b4-4271-a936-3795037adc6e","Type":"ContainerStarted","Data":"0b71e057eb238d3dbb1ab2e0614fc6c5c93109729f9cb590b51da8c9ddfc9452"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.686287 4861 generic.go:334] "Generic (PLEG): container finished" podID="009b6cf3-3e73-4b27-b249-1c72de7bd3dd" containerID="8760f9ad8d6b335e89e2563ea6965053f4584a6421aface8389cc41bb90c0dca" exitCode=0 Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.686358 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" event={"ID":"009b6cf3-3e73-4b27-b249-1c72de7bd3dd","Type":"ContainerDied","Data":"8760f9ad8d6b335e89e2563ea6965053f4584a6421aface8389cc41bb90c0dca"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.686402 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" event={"ID":"009b6cf3-3e73-4b27-b249-1c72de7bd3dd","Type":"ContainerStarted","Data":"52ff1225d2585860d78a928c5d2708f891832369fa028120510cd2340994352a"} Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.689665 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm5wt\" (UniqueName: \"kubernetes.io/projected/432faa21-42d4-454a-a36f-dcf427e79e96-kube-api-access-cm5wt\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.729914 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-config\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.729971 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730025 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8h5m\" (UniqueName: \"kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730067 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730114 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhr8f\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730399 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w9wb\" (UniqueName: \"kubernetes.io/projected/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-kube-api-access-7w9wb\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730466 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ffaef4-ac61-4166-bf7f-4135a0a65934-config\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730503 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-images\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.730549 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.230526677 +0000 UTC m=+253.542724210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.730702 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8ffaef4-ac61-4166-bf7f-4135a0a65934-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.731698 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.731906 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733141 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h52wb\" (UniqueName: \"kubernetes.io/projected/58100d65-ee1f-4e57-9140-5519af53b4c6-kube-api-access-h52wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733361 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733390 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ffaef4-ac61-4166-bf7f-4135a0a65934-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733472 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-trusted-ca\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733600 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l64dc\" (UniqueName: \"kubernetes.io/projected/def607d8-9395-4eab-af6b-a03476b186ed-kube-api-access-l64dc\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733636 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733692 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78jnd\" (UniqueName: \"kubernetes.io/projected/7f160fb8-7a29-4f18-a072-ac62270476de-kube-api-access-78jnd\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733713 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.733731 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734003 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734088 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/58100d65-ee1f-4e57-9140-5519af53b4c6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734398 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f160fb8-7a29-4f18-a072-ac62270476de-serving-cert\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734473 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734606 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/def607d8-9395-4eab-af6b-a03476b186ed-proxy-tls\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.734645 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.758608 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558890-7spp4"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.765631 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.769864 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.835253 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.835516 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.335475474 +0000 UTC m=+253.647673017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836180 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-csi-data-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836246 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836308 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/58100d65-ee1f-4e57-9140-5519af53b4c6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836333 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836398 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/53dc9c56-f3df-43eb-a8af-f068904d8ab0-cert\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836442 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836469 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f160fb8-7a29-4f18-a072-ac62270476de-serving-cert\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836486 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836531 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836611 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-node-bootstrap-token\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836662 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/def607d8-9395-4eab-af6b-a03476b186ed-proxy-tls\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836683 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836701 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836719 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7d5e0358-4756-4e8c-aec1-baf3f5d52699-metrics-tls\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836756 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836783 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-config\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836802 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836843 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836859 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5rqh\" (UniqueName: \"kubernetes.io/projected/53dc9c56-f3df-43eb-a8af-f068904d8ab0-kube-api-access-g5rqh\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836890 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7277\" (UniqueName: \"kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836910 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8h5m\" (UniqueName: \"kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836950 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.836994 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-certs\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837032 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhr8f\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837068 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv2wq\" (UniqueName: \"kubernetes.io/projected/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-kube-api-access-nv2wq\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837088 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837108 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837127 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837198 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w9wb\" (UniqueName: \"kubernetes.io/projected/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-kube-api-access-7w9wb\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837236 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837259 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ffaef4-ac61-4166-bf7f-4135a0a65934-config\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837282 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837313 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-images\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837330 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-socket-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837359 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837397 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837414 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8ffaef4-ac61-4166-bf7f-4135a0a65934-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837452 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d5e0358-4756-4e8c-aec1-baf3f5d52699-config-volume\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837483 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837500 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg87d\" (UniqueName: \"kubernetes.io/projected/63356f73-2c6e-472c-a38b-a431c562d4e3-kube-api-access-pg87d\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837526 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837608 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h52wb\" (UniqueName: \"kubernetes.io/projected/58100d65-ee1f-4e57-9140-5519af53b4c6-kube-api-access-h52wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837642 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837677 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837717 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837745 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ffaef4-ac61-4166-bf7f-4135a0a65934-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837807 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837837 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837857 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837874 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-trusted-ca\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837892 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l64dc\" (UniqueName: \"kubernetes.io/projected/def607d8-9395-4eab-af6b-a03476b186ed-kube-api-access-l64dc\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837911 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-plugins-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837948 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-registration-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.837986 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838005 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78jnd\" (UniqueName: \"kubernetes.io/projected/7f160fb8-7a29-4f18-a072-ac62270476de-kube-api-access-78jnd\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838039 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838057 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838095 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838113 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838190 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838214 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838273 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838290 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838306 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-mountpoint-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838323 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.838339 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6r5\" (UniqueName: \"kubernetes.io/projected/7d5e0358-4756-4e8c-aec1-baf3f5d52699-kube-api-access-9n6r5\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.842776 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-config\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.843897 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.845052 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.845802 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.846075 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.850100 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.851360 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3589ff40-d99e-4598-8f93-b8721775f117-metrics-tls\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.851763 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.852124 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.852695 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6ef26501-9849-4bd4-8976-38cc771e868e-images\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.852825 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.854389 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-service-ca\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.860944 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/58100d65-ee1f-4e57-9140-5519af53b4c6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.861626 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.861910 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.862293 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f160fb8-7a29-4f18-a072-ac62270476de-serving-cert\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.863090 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.866008 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/def607d8-9395-4eab-af6b-a03476b186ed-proxy-tls\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.866193 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.866254 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.875067 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.875961 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.878067 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h52wb\" (UniqueName: \"kubernetes.io/projected/58100d65-ee1f-4e57-9140-5519af53b4c6-kube-api-access-h52wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-g2qt2\" (UID: \"58100d65-ee1f-4e57-9140-5519af53b4c6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.886326 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adba3972-5deb-41ab-adcf-53795715f713-config\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.886926 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-294fg"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.887686 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.891149 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-trusted-ca\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.891439 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: W0315 00:10:59.891615 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28681c01_d949_429f_ad5c_6084f6d30e81.slice/crio-9218dfaa242f0d97a1041dc555d80af76e9c0322033a3b899335e946f67653a9 WatchSource:0}: Error finding container 9218dfaa242f0d97a1041dc555d80af76e9c0322033a3b899335e946f67653a9: Status 404 returned error can't find the container with id 9218dfaa242f0d97a1041dc555d80af76e9c0322033a3b899335e946f67653a9 Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.892459 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ef26501-9849-4bd4-8976-38cc771e868e-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fbnkg\" (UID: \"6ef26501-9849-4bd4-8976-38cc771e868e\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.892721 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.892920 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.893493 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-oauth-serving-cert\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.894423 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.895477 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.897600 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0c9069d-5138-4fd9-9c2f-78b5224287e8-trusted-ca-bundle\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.897924 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.39790587 +0000 UTC m=+253.710103393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.898210 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f160fb8-7a29-4f18-a072-ac62270476de-config\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.901323 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.901998 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/def607d8-9395-4eab-af6b-a03476b186ed-images\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.903019 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ffaef4-ac61-4166-bf7f-4135a0a65934-config\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.903134 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ffaef4-ac61-4166-bf7f-4135a0a65934-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.907000 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7277\" (UniqueName: \"kubernetes.io/projected/3589ff40-d99e-4598-8f93-b8721775f117-kube-api-access-p7277\") pod \"dns-operator-744455d44c-lns4z\" (UID: \"3589ff40-d99e-4598-8f93-b8721775f117\") " pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.907282 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.909751 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-z2gvn"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.910671 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b0c9069d-5138-4fd9-9c2f-78b5224287e8-console-oauth-config\") pod \"console-f9d7485db-4c7wr\" (UID: \"b0c9069d-5138-4fd9-9c2f-78b5224287e8\") " pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.911121 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/432faa21-42d4-454a-a36f-dcf427e79e96-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-7jkmx\" (UID: \"432faa21-42d4-454a-a36f-dcf427e79e96\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.919058 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.924658 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/adba3972-5deb-41ab-adcf-53795715f713-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-kmj2d\" (UID: \"adba3972-5deb-41ab-adcf-53795715f713\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.926444 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.926448 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8ffaef4-ac61-4166-bf7f-4135a0a65934-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hn2vr\" (UID: \"a8ffaef4-ac61-4166-bf7f-4135a0a65934\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.932739 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j2krj\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.938909 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939261 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-mountpoint-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939289 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6r5\" (UniqueName: \"kubernetes.io/projected/7d5e0358-4756-4e8c-aec1-baf3f5d52699-kube-api-access-9n6r5\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939308 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-csi-data-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939332 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/53dc9c56-f3df-43eb-a8af-f068904d8ab0-cert\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939355 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-node-bootstrap-token\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939380 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7d5e0358-4756-4e8c-aec1-baf3f5d52699-metrics-tls\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939405 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5rqh\" (UniqueName: \"kubernetes.io/projected/53dc9c56-f3df-43eb-a8af-f068904d8ab0-kube-api-access-g5rqh\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939442 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-certs\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939464 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv2wq\" (UniqueName: \"kubernetes.io/projected/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-kube-api-access-nv2wq\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939499 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-socket-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939530 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d5e0358-4756-4e8c-aec1-baf3f5d52699-config-volume\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939548 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg87d\" (UniqueName: \"kubernetes.io/projected/63356f73-2c6e-472c-a38b-a431c562d4e3-kube-api-access-pg87d\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939618 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-plugins-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939647 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-registration-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.939914 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-registration-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.940067 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.941387 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.943534 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.947056 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-socket-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.947174 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-csi-data-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.947239 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-mountpoint-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.947715 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d5e0358-4756-4e8c-aec1-baf3f5d52699-config-volume\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.947818 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-plugins-dir\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.950099 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7d5e0358-4756-4e8c-aec1-baf3f5d52699-metrics-tls\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:10:59 crc kubenswrapper[4861]: E0315 00:10:59.950611 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.450590724 +0000 UTC m=+253.762788257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.953257 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhr8f\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.953967 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.954794 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.955717 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-phl6h"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.956019 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db"] Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.962312 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/53dc9c56-f3df-43eb-a8af-f068904d8ab0-cert\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.966414 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-certs\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.969219 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78jnd\" (UniqueName: \"kubernetes.io/projected/7f160fb8-7a29-4f18-a072-ac62270476de-kube-api-access-78jnd\") pod \"console-operator-58897d9998-k9d4t\" (UID: \"7f160fb8-7a29-4f18-a072-ac62270476de\") " pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.972609 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/63356f73-2c6e-472c-a38b-a431c562d4e3-node-bootstrap-token\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.978869 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w9wb\" (UniqueName: \"kubernetes.io/projected/f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7-kube-api-access-7w9wb\") pod \"package-server-manager-789f6589d5-hcv9p\" (UID: \"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.979202 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:10:59 crc kubenswrapper[4861]: I0315 00:10:59.997042 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.001031 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8h5m\" (UniqueName: \"kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m\") pod \"marketplace-operator-79b997595-kt6gz\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.025333 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.027497 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l64dc\" (UniqueName: \"kubernetes.io/projected/def607d8-9395-4eab-af6b-a03476b186ed-kube-api-access-l64dc\") pod \"machine-config-operator-74547568cd-5c5ms\" (UID: \"def607d8-9395-4eab-af6b-a03476b186ed\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:11:00 crc kubenswrapper[4861]: W0315 00:11:00.027601 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1901c2b3_404c_4363_a77d_724f71aedb02.slice/crio-6d6ce2729894b1a268e5a404e503b219dfa78a0b4b3935350b394006f1670543 WatchSource:0}: Error finding container 6d6ce2729894b1a268e5a404e503b219dfa78a0b4b3935350b394006f1670543: Status 404 returned error can't find the container with id 6d6ce2729894b1a268e5a404e503b219dfa78a0b4b3935350b394006f1670543 Mar 15 00:11:00 crc kubenswrapper[4861]: W0315 00:11:00.030512 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod164464a6_d9c6_475e_a7ae_707365c5ec51.slice/crio-d4249d9c8d23b5b753a6cde41582b5e573db4b2095ae6a0dfbb5e9c11ef9df84 WatchSource:0}: Error finding container d4249d9c8d23b5b753a6cde41582b5e573db4b2095ae6a0dfbb5e9c11ef9df84: Status 404 returned error can't find the container with id d4249d9c8d23b5b753a6cde41582b5e573db4b2095ae6a0dfbb5e9c11ef9df84 Mar 15 00:11:00 crc kubenswrapper[4861]: W0315 00:11:00.034025 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7447a1d1_672c_4397_99ac_3657c4586eb3.slice/crio-7d271436788ced2b8a789614f8ac4ce249864a7d227b6c0a9bc2e66d4e073217 WatchSource:0}: Error finding container 7d271436788ced2b8a789614f8ac4ce249864a7d227b6c0a9bc2e66d4e073217: Status 404 returned error can't find the container with id 7d271436788ced2b8a789614f8ac4ce249864a7d227b6c0a9bc2e66d4e073217 Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.040618 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.045233 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.045477 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.545462841 +0000 UTC m=+253.857660374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.074952 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5rqh\" (UniqueName: \"kubernetes.io/projected/53dc9c56-f3df-43eb-a8af-f068904d8ab0-kube-api-access-g5rqh\") pod \"ingress-canary-zqb9z\" (UID: \"53dc9c56-f3df-43eb-a8af-f068904d8ab0\") " pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.100508 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6r5\" (UniqueName: \"kubernetes.io/projected/7d5e0358-4756-4e8c-aec1-baf3f5d52699-kube-api-access-9n6r5\") pod \"dns-default-k6lv8\" (UID: \"7d5e0358-4756-4e8c-aec1-baf3f5d52699\") " pod="openshift-dns/dns-default-k6lv8" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.104051 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.112920 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.120448 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.120687 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv2wq\" (UniqueName: \"kubernetes.io/projected/1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1-kube-api-access-nv2wq\") pod \"csi-hostpathplugin-7sdg6\" (UID: \"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.131083 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.141331 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.141871 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.641851779 +0000 UTC m=+253.954049312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.156217 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg87d\" (UniqueName: \"kubernetes.io/projected/63356f73-2c6e-472c-a38b-a431c562d4e3-kube-api-access-pg87d\") pod \"machine-config-server-hjq5h\" (UID: \"63356f73-2c6e-472c-a38b-a431c562d4e3\") " pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.178454 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hjq5h" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.201032 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.208989 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-zqb9z" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.214451 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-k6lv8" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.251069 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.255150 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.755120749 +0000 UTC m=+254.067318282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.288056 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.352760 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.353165 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.853145591 +0000 UTC m=+254.165343124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.385028 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.413814 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.454706 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.455128 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:00.955113098 +0000 UTC m=+254.267310631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.555860 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.556237 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.056211152 +0000 UTC m=+254.368408685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.557177 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.557636 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.05762225 +0000 UTC m=+254.369819783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.661638 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.662073 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.162052564 +0000 UTC m=+254.474250097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.669664 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4c7wr"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.704350 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j2krj"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.711247 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fbnkg"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.733697 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.737294 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-k9d4t"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.752731 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" event={"ID":"5dd65149-6fb3-492b-9719-e5e22e51eedd","Type":"ContainerStarted","Data":"99c0c3b0bea7321882d076e26ea4e41ca0d6a7e4f8d976e84402969d658c6f39"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.770430 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.771261 4861 generic.go:334] "Generic (PLEG): container finished" podID="000fdf2c-16b4-4271-a936-3795037adc6e" containerID="75d3a4f7a1ad0686f27bad2e174b4198c9b7d850f499c70a19d70e222aa82a61" exitCode=0 Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.771332 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" event={"ID":"000fdf2c-16b4-4271-a936-3795037adc6e","Type":"ContainerDied","Data":"75d3a4f7a1ad0686f27bad2e174b4198c9b7d850f499c70a19d70e222aa82a61"} Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.771499 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.271482741 +0000 UTC m=+254.583680274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.806215 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" event={"ID":"5569250f-d178-41af-ac06-64aa34373900","Type":"ContainerStarted","Data":"347d6d9ce99d8d658c723e294303bc9432cbc4bd5c45012101e41e6c99533b67"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.806730 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.816929 4861 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rmkcs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.817018 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" podUID="5569250f-d178-41af-ac06-64aa34373900" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.824090 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" event={"ID":"8fbab285-e809-4147-846b-bd2252bf4f18","Type":"ContainerStarted","Data":"732ed14ed809fc9a94c6b627fc97a6a2fbc347d87589d30befedbeaac3d79e27"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.825199 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.840163 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" event={"ID":"164464a6-d9c6-475e-a7ae-707365c5ec51","Type":"ContainerStarted","Data":"d4249d9c8d23b5b753a6cde41582b5e573db4b2095ae6a0dfbb5e9c11ef9df84"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.863829 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" event={"ID":"0a957ae9-6295-4228-9dfc-6f50cdfef5e2","Type":"ContainerStarted","Data":"cf7c425a0c0da562583ed9f4a88f3409d0407391c7b261e8242151f2a3ffb3a2"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.866917 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" event={"ID":"a861f914-a0a6-4b7e-98b7-94cbc3ea942d","Type":"ContainerStarted","Data":"babe39af81f56f79fccb12518973fbb61ce7fa9cc8aee8efb0e837f649288171"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.872243 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.872800 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.37275978 +0000 UTC m=+254.684957313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.903868 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2"] Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.930605 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" event={"ID":"a71a5c20-c130-4d0d-9bef-f4c12204ae67","Type":"ContainerStarted","Data":"c6a75d282a7797442ea2fd3e606d8e15bd2821cdc074734252f5c9b7e1d4f49b"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.946617 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" event={"ID":"43636426-c477-4b47-b209-87df8d3414bc","Type":"ContainerStarted","Data":"9a10a7796bb2a5782b15bac18dbdcc6df87767f6dc0ae3397110c9691dae6a7b"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.947491 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.960538 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" event={"ID":"0f7ab638-daa2-4d39-80e7-6a8061f62357","Type":"ContainerStarted","Data":"eadf32d4a71a3c298ce817968ddcbc7cdfeb43cf4a3e74af6b4eb478d748dddd"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.960607 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" event={"ID":"0f7ab638-daa2-4d39-80e7-6a8061f62357","Type":"ContainerStarted","Data":"234675c2449244f913ddd5278ed779b1d7c9d2c8e1aee292f5518ea1d8682812"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.972009 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" event={"ID":"a2b9027a-9381-417c-b500-16822f802828","Type":"ContainerStarted","Data":"4c65a44d5cb98d06f7eaa42d6e203ea0101f7c43d4eb05d1d1bc0c21f1d9b3f4"} Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.973472 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:00 crc kubenswrapper[4861]: E0315 00:11:00.975433 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.475415965 +0000 UTC m=+254.787613678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.981151 4861 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7hrpb container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.981649 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" podUID="43636426-c477-4b47-b209-87df8d3414bc" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Mar 15 00:11:00 crc kubenswrapper[4861]: I0315 00:11:00.990776 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" event={"ID":"28681c01-d949-429f-ad5c-6084f6d30e81","Type":"ContainerStarted","Data":"9218dfaa242f0d97a1041dc555d80af76e9c0322033a3b899335e946f67653a9"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.004895 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558890-7spp4" event={"ID":"4585669d-fc84-4c80-9cc3-0185b7fb9ce0","Type":"ContainerStarted","Data":"c63ca3f0ec4db37a08dae1b2ac77bfdb8b254e622c0e466843fef8aa3c420313"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.009574 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" event={"ID":"009b6cf3-3e73-4b27-b249-1c72de7bd3dd","Type":"ContainerStarted","Data":"fdcbb88a3c840c53492fc0b55de32e061c9e4625b1df8458605db2c3e1e96aae"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.011288 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" event={"ID":"1901c2b3-404c-4363-a77d-724f71aedb02","Type":"ContainerStarted","Data":"6d6ce2729894b1a268e5a404e503b219dfa78a0b4b3935350b394006f1670543"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.022004 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" event={"ID":"c6174b71-a5cc-4a0e-902b-9a9091d15d83","Type":"ContainerStarted","Data":"eb7478619d91d7ecec547ea0be38f3ba57c10422543d246fe82e38ac47837247"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.022778 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.024628 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" event={"ID":"59b39d47-b227-436a-9437-688c54bd4e05","Type":"ContainerStarted","Data":"e0e7f47d3c1424256c55a005e859a836b993788ffe7e8e0291df92dd838e856a"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.026737 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" event={"ID":"381e7ca1-1798-42ef-8df1-bc0d15f88e23","Type":"ContainerStarted","Data":"7f7ccd7c95be0f2f35a5a6e878499b65728ec6bd9df54f048a2b0fd681c6dcbf"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.028973 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.036889 4861 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-5fg69 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.036953 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.041763 4861 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bk6w5 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.041827 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" podUID="381e7ca1-1798-42ef-8df1-bc0d15f88e23" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.058701 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" event={"ID":"997eb245-8ce2-4eba-a167-444b69f138c2","Type":"ContainerStarted","Data":"9de61e69c70c74f5f58cb8ea9b9636d795bb7521e03ec0c58f676562c862cb41"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.075527 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" event={"ID":"7447a1d1-672c-4397-99ac-3657c4586eb3","Type":"ContainerStarted","Data":"7d271436788ced2b8a789614f8ac4ce249864a7d227b6c0a9bc2e66d4e073217"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.081150 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.085297 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.585244914 +0000 UTC m=+254.897442447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.086935 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" event={"ID":"b1bcc8c9-7f94-4733-b691-53f0e30df85c","Type":"ContainerStarted","Data":"b777eebf5bdc12efded8e0bcd5b267b659557080f68947e61ba8bce73cfc5239"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.088874 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.091970 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.591948644 +0000 UTC m=+254.904146177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.095520 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" event={"ID":"8e903079-fff6-462b-a805-3dac5d18a169","Type":"ContainerStarted","Data":"c1e3b0cafa12a41bc934a8490d8dcad44c407965f093c76eec629a57d7250abb"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.108706 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" event={"ID":"09c6e2b5-3909-400a-a052-5e540226011b","Type":"ContainerStarted","Data":"ce6f298dd3aed453172fe8984b98b9746c4c2b45dedc30236bdeee2faa47afaa"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.108781 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" event={"ID":"09c6e2b5-3909-400a-a052-5e540226011b","Type":"ContainerStarted","Data":"4ddc89b721b25f924a97523b84a0de3b43358b0a4ee0626962a474c4aca84b2d"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.120166 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lgbxx" event={"ID":"631708f5-e5a1-4d6f-8858-60ea4de43cd9","Type":"ContainerStarted","Data":"ddbced5642524cf7ebb2d1c1e888172b5854e0fb94093cc41f9ebe005f38d836"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.167501 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" event={"ID":"116d02ff-db52-4e72-92a5-9b2de33db24a","Type":"ContainerStarted","Data":"240cf74741859c07c05b07ae78a2d235c9ba4e982d7b8334b44b145065bf81fb"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.179017 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" event={"ID":"a8ffaef4-ac61-4166-bf7f-4135a0a65934","Type":"ContainerStarted","Data":"375c2520b8a503e70dcf83cdc183e822afd4adac88c55dbed8b81dd18b81cb89"} Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.180884 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.180953 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.191910 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.192616 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.692546674 +0000 UTC m=+255.004744207 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.193397 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.196484 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.696468479 +0000 UTC m=+255.008666002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.294881 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.295160 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.795128018 +0000 UTC m=+255.107325551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.295620 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.297150 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.797134692 +0000 UTC m=+255.109332225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.300080 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29558880-xw69x" podStartSLOduration=192.30004934 podStartE2EDuration="3m12.30004934s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:01.294384858 +0000 UTC m=+254.606582401" watchObservedRunningTime="2026-03-15 00:11:01.30004934 +0000 UTC m=+254.612246873" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.313007 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p"] Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.331646 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-lns4z"] Mar 15 00:11:01 crc kubenswrapper[4861]: W0315 00:11:01.366250 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3589ff40_d99e_4598_8f93_b8721775f117.slice/crio-03e55a0a36ae2ed4a00d2c89fa8e2d1a4f9b3c8ad66010723e5096a5d31fa02f WatchSource:0}: Error finding container 03e55a0a36ae2ed4a00d2c89fa8e2d1a4f9b3c8ad66010723e5096a5d31fa02f: Status 404 returned error can't find the container with id 03e55a0a36ae2ed4a00d2c89fa8e2d1a4f9b3c8ad66010723e5096a5d31fa02f Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.401263 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.402098 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.902054518 +0000 UTC m=+255.214252051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.402937 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.403639 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:01.9036288 +0000 UTC m=+255.215826333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.514967 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.515398 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.01538122 +0000 UTC m=+255.327578753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.522978 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sdg6"] Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.613618 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-k6lv8"] Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.644462 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.687026 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.186996298 +0000 UTC m=+255.499193831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.710360 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-zqb9z"] Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.745489 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms"] Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.756432 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.757188 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.257164351 +0000 UTC m=+255.569361904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.782995 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.799374 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xzvqd" podStartSLOduration=192.799333613 podStartE2EDuration="3m12.799333613s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:01.798938662 +0000 UTC m=+255.111136205" watchObservedRunningTime="2026-03-15 00:11:01.799333613 +0000 UTC m=+255.111531146" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.814155 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:01 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:01 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:01 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.814235 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.859010 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.859381 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.359365995 +0000 UTC m=+255.671563528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.931109 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" podStartSLOduration=191.93108699 podStartE2EDuration="3m11.93108699s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:01.923479155 +0000 UTC m=+255.235676688" watchObservedRunningTime="2026-03-15 00:11:01.93108699 +0000 UTC m=+255.243284523" Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.960359 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:01 crc kubenswrapper[4861]: E0315 00:11:01.961152 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.461133386 +0000 UTC m=+255.773330919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:01 crc kubenswrapper[4861]: I0315 00:11:01.972449 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-lvzfq" podStartSLOduration=192.972422719 podStartE2EDuration="3m12.972422719s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:01.969692196 +0000 UTC m=+255.281889749" watchObservedRunningTime="2026-03-15 00:11:01.972422719 +0000 UTC m=+255.284620252" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.088768 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.089212 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.589196624 +0000 UTC m=+255.901394157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.090576 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" podStartSLOduration=192.090549701 podStartE2EDuration="3m12.090549701s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.088210348 +0000 UTC m=+255.400407881" watchObservedRunningTime="2026-03-15 00:11:02.090549701 +0000 UTC m=+255.402747224" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.193736 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.194362 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.694338027 +0000 UTC m=+256.006535560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.265357 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" event={"ID":"000fdf2c-16b4-4271-a936-3795037adc6e","Type":"ContainerStarted","Data":"0b6222898c1974f02c267eb2d26a86f7d4dfb4709f4e070faeb385d244d40934"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.265437 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.272247 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-s9k7b" podStartSLOduration=192.272222358 podStartE2EDuration="3m12.272222358s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.201884279 +0000 UTC m=+255.514081812" watchObservedRunningTime="2026-03-15 00:11:02.272222358 +0000 UTC m=+255.584419911" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.288701 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.288770 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.296716 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.297211 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.797195747 +0000 UTC m=+256.109393280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.335369 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8pv8j" podStartSLOduration=193.335344782 podStartE2EDuration="3m13.335344782s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.333997066 +0000 UTC m=+255.646194609" watchObservedRunningTime="2026-03-15 00:11:02.335344782 +0000 UTC m=+255.647542315" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.341904 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" podStartSLOduration=192.341887477 podStartE2EDuration="3m12.341887477s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.276855452 +0000 UTC m=+255.589053015" watchObservedRunningTime="2026-03-15 00:11:02.341887477 +0000 UTC m=+255.654085010" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.364451 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" event={"ID":"a2b9027a-9381-417c-b500-16822f802828","Type":"ContainerStarted","Data":"77c0b9893b80a00c4211fd8a9bad32355904e6c5e6118d9f513d34804d9a0400"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.378205 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-dbghq" podStartSLOduration=193.378178532 podStartE2EDuration="3m13.378178532s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.375643813 +0000 UTC m=+255.687841356" watchObservedRunningTime="2026-03-15 00:11:02.378178532 +0000 UTC m=+255.690376065" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.403280 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.403806 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:02.903774168 +0000 UTC m=+256.215971701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.406050 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4c7wr" event={"ID":"b0c9069d-5138-4fd9-9c2f-78b5224287e8","Type":"ContainerStarted","Data":"5b35216e5f17447e9abdce73821de5fa573b211bc716b2c5fd4402982e066a2d"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.406104 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4c7wr" event={"ID":"b0c9069d-5138-4fd9-9c2f-78b5224287e8","Type":"ContainerStarted","Data":"deff286308631f44474a77dead28d42ab8fba95795ab6be8704ddd13618c6ddf"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.430006 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" event={"ID":"def607d8-9395-4eab-af6b-a03476b186ed","Type":"ContainerStarted","Data":"0e0f95aec9d8239d2bd08cc6d7d0d26f879b008a1666aac8ca72cfee48bb76ee"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.453828 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" podStartSLOduration=193.453801892 podStartE2EDuration="3m13.453801892s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.44816951 +0000 UTC m=+255.760367063" watchObservedRunningTime="2026-03-15 00:11:02.453801892 +0000 UTC m=+255.765999425" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.495631 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53150: no serving certificate available for the kubelet" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.502427 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" podStartSLOduration=192.502408207 podStartE2EDuration="3m12.502408207s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.496314103 +0000 UTC m=+255.808511636" watchObservedRunningTime="2026-03-15 00:11:02.502408207 +0000 UTC m=+255.814605740" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.503099 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" event={"ID":"164464a6-d9c6-475e-a7ae-707365c5ec51","Type":"ContainerStarted","Data":"cbdb716411bca1363c0d2eb2ceb3ee46fcb3aa63a3b5e120d762751a0261e14d"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.505632 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.507346 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.007331108 +0000 UTC m=+256.319528641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.570314 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" event={"ID":"a861f914-a0a6-4b7e-98b7-94cbc3ea942d","Type":"ContainerStarted","Data":"e76c5511eb5cc593c083297494e8a01e24a08cef322d7809ed0c70990c01a8fe"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.573541 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" podStartSLOduration=192.573527676 podStartE2EDuration="3m12.573527676s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.528035844 +0000 UTC m=+255.840233377" watchObservedRunningTime="2026-03-15 00:11:02.573527676 +0000 UTC m=+255.885725209" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.596973 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-4h9nq" podStartSLOduration=193.596945624 podStartE2EDuration="3m13.596945624s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.571994495 +0000 UTC m=+255.884192028" watchObservedRunningTime="2026-03-15 00:11:02.596945624 +0000 UTC m=+255.909143177" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.608505 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53156: no serving certificate available for the kubelet" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.613227 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.615027 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.114976408 +0000 UTC m=+256.427173941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.615981 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lgbxx" podStartSLOduration=193.615953755 podStartE2EDuration="3m13.615953755s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.601585089 +0000 UTC m=+255.913782632" watchObservedRunningTime="2026-03-15 00:11:02.615953755 +0000 UTC m=+255.928151278" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.622578 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.626687 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.126653232 +0000 UTC m=+256.438850765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.628675 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" event={"ID":"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7","Type":"ContainerStarted","Data":"4b047653c8366ae4e02a2e04f36dcd14092e86d50b9eb421464a54f9b0864b4f"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.685001 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53158: no serving certificate available for the kubelet" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.695057 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-z2gvn" podStartSLOduration=192.695033887 podStartE2EDuration="3m12.695033887s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.694666667 +0000 UTC m=+256.006864200" watchObservedRunningTime="2026-03-15 00:11:02.695033887 +0000 UTC m=+256.007231420" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.707090 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" podStartSLOduration=193.70706186 podStartE2EDuration="3m13.70706186s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.63887449 +0000 UTC m=+255.951072023" watchObservedRunningTime="2026-03-15 00:11:02.70706186 +0000 UTC m=+256.019259393" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.734697 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.735342 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.736704 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.236680215 +0000 UTC m=+256.548877748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.739008 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4c7wr" podStartSLOduration=193.738985697 podStartE2EDuration="3m13.738985697s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.734052825 +0000 UTC m=+256.046250358" watchObservedRunningTime="2026-03-15 00:11:02.738985697 +0000 UTC m=+256.051183230" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.756441 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" event={"ID":"009b6cf3-3e73-4b27-b249-1c72de7bd3dd","Type":"ContainerStarted","Data":"11950cb3ea215c359f3bc779aba14af27578a36deb9cec75b78c18265602ef9d"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.759498 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" event={"ID":"7f160fb8-7a29-4f18-a072-ac62270476de","Type":"ContainerStarted","Data":"b347cc0097e56538c4e0f118566cdbee88960f0f6d834f3c7178288404d0ff87"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.760041 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.762933 4861 patch_prober.go:28] interesting pod/console-operator-58897d9998-k9d4t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.762996 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" podUID="7f160fb8-7a29-4f18-a072-ac62270476de" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.792072 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53166: no serving certificate available for the kubelet" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.793128 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" event={"ID":"a8ffaef4-ac61-4166-bf7f-4135a0a65934","Type":"ContainerStarted","Data":"62d06a027934076744a7f2110f2fc5bf37c8f385fa684b2e347751842cb0ce06"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.794227 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-phl6h" podStartSLOduration=193.794207139 podStartE2EDuration="3m13.794207139s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.790739527 +0000 UTC m=+256.102937060" watchObservedRunningTime="2026-03-15 00:11:02.794207139 +0000 UTC m=+256.106404672" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.803057 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:02 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:02 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:02 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.803147 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.809781 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k6lv8" event={"ID":"7d5e0358-4756-4e8c-aec1-baf3f5d52699","Type":"ContainerStarted","Data":"042329db23e981fdb2e8e2927ec15e0ce678b9130816fbb61d57d7eb98a8fefa"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.819584 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" event={"ID":"6ef26501-9849-4bd4-8976-38cc771e868e","Type":"ContainerStarted","Data":"bf6e4565897f35964ac3287704339ad9c6a4498ccd47eece71816c9167988017"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.819635 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" event={"ID":"6ef26501-9849-4bd4-8976-38cc771e868e","Type":"ContainerStarted","Data":"b448af965687470ce21eace4486fe54965bb49a721d5cb76a3047c2b49f5b792"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.847784 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.867115 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.367091146 +0000 UTC m=+256.679288679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.878310 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" podStartSLOduration=193.878284216 podStartE2EDuration="3m13.878284216s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.82926203 +0000 UTC m=+256.141459563" watchObservedRunningTime="2026-03-15 00:11:02.878284216 +0000 UTC m=+256.190481749" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.892273 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" event={"ID":"997eb245-8ce2-4eba-a167-444b69f138c2","Type":"ContainerStarted","Data":"9a11dfe6a4a5b85cad7dbfd633d2bae4ed21beabf87e51a3745072260adad681"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.917066 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" event={"ID":"8fbab285-e809-4147-846b-bd2252bf4f18","Type":"ContainerStarted","Data":"0e7c3f21a263dd049cf851bcd60ebb791fbe04a6860e84d062a46a496bd26aee"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.942332 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" event={"ID":"8e903079-fff6-462b-a805-3dac5d18a169","Type":"ContainerStarted","Data":"d368b5ee3b476d7aa06f95f6b0df4d0068bb83b17b4f583738481f9aaff8680e"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.949077 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" podStartSLOduration=193.949057917 podStartE2EDuration="3m13.949057917s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.88212779 +0000 UTC m=+256.194325333" watchObservedRunningTime="2026-03-15 00:11:02.949057917 +0000 UTC m=+256.261255450" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.952001 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:02 crc kubenswrapper[4861]: E0315 00:11:02.955182 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.45515677 +0000 UTC m=+256.767354293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.977174 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zqb9z" event={"ID":"53dc9c56-f3df-43eb-a8af-f068904d8ab0","Type":"ContainerStarted","Data":"42d3488cc892838349a05627480208ab1b189cec46b5a7140502cf71492f1c1a"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.979850 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hn2vr" podStartSLOduration=193.979825823 podStartE2EDuration="3m13.979825823s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.976677458 +0000 UTC m=+256.288874991" watchObservedRunningTime="2026-03-15 00:11:02.979825823 +0000 UTC m=+256.292023356" Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.992777 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" event={"ID":"adba3972-5deb-41ab-adcf-53795715f713","Type":"ContainerStarted","Data":"8d4585f16c341567a88b6560b14b847469a4649ca90aad1b3b15e5b87c4c53cb"} Mar 15 00:11:02 crc kubenswrapper[4861]: I0315 00:11:02.992848 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" event={"ID":"adba3972-5deb-41ab-adcf-53795715f713","Type":"ContainerStarted","Data":"60e227d1bd2a39f67ac42c2d65ca2edff1d77f60b4f4901aa7a019a06c698b2f"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.011197 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53174: no serving certificate available for the kubelet" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.013589 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-p96db" podStartSLOduration=194.008527903 podStartE2EDuration="3m14.008527903s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:02.995785911 +0000 UTC m=+256.307983454" watchObservedRunningTime="2026-03-15 00:11:03.008527903 +0000 UTC m=+256.320725436" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.015531 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" event={"ID":"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1","Type":"ContainerStarted","Data":"18a998e6865ac56ef792f67ea4db233db9b37ee433f66ef81374292b2210fc61"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.031246 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerStarted","Data":"7fa5d6ee90d0d1e86542acbda0ec6ef580bc847d471f1e06e1ddeee829b2acf9"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.031902 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.037379 4861 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kt6gz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.037446 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.049309 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vfvmf" podStartSLOduration=194.049281817 podStartE2EDuration="3m14.049281817s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.049106562 +0000 UTC m=+256.361304085" watchObservedRunningTime="2026-03-15 00:11:03.049281817 +0000 UTC m=+256.361479350" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.055072 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.059785 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.559762979 +0000 UTC m=+256.871960512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.074927 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" event={"ID":"7447a1d1-672c-4397-99ac-3657c4586eb3","Type":"ContainerStarted","Data":"c1e25c2a27e672103ee399a556a6e78307209eca5395238024e57e6bbd727408"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.076045 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.092013 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" event={"ID":"1e8992d1-2a3f-4a3b-b857-821cfb8bed31","Type":"ContainerStarted","Data":"cb6ebb91900b2fad5eec9fcfeb9e366b28505cf7954d8169e2313ffebd95d69b"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.093178 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.096390 4861 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-j2krj container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" start-of-body= Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.096433 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.099795 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.110915 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8chcv" podStartSLOduration=194.110897972 podStartE2EDuration="3m14.110897972s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.105061354 +0000 UTC m=+256.417258887" watchObservedRunningTime="2026-03-15 00:11:03.110897972 +0000 UTC m=+256.423095505" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.117154 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" event={"ID":"58100d65-ee1f-4e57-9140-5519af53b4c6","Type":"ContainerStarted","Data":"41b9abbf9c0375300b47c594c0e84065b9d2858aa9561ca7a2fd4688e88a5a91"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.117191 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" event={"ID":"58100d65-ee1f-4e57-9140-5519af53b4c6","Type":"ContainerStarted","Data":"5cfdd87b51a0eabb066950a1381c83a2b65859259cba33782f09cfb9236fd9ba"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.142212 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" podStartSLOduration=193.142186991 podStartE2EDuration="3m13.142186991s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.134665499 +0000 UTC m=+256.446863032" watchObservedRunningTime="2026-03-15 00:11:03.142186991 +0000 UTC m=+256.454384524" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.157093 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.158018 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.162299 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.164427 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" event={"ID":"3589ff40-d99e-4598-8f93-b8721775f117","Type":"ContainerStarted","Data":"03e55a0a36ae2ed4a00d2c89fa8e2d1a4f9b3c8ad66010723e5096a5d31fa02f"} Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.166083 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.666045442 +0000 UTC m=+256.978242975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.178925 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53182: no serving certificate available for the kubelet" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.224887 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" event={"ID":"432faa21-42d4-454a-a36f-dcf427e79e96","Type":"ContainerStarted","Data":"b5b8b76ee5b2bd6be22f0438d54b678b20e69928bdee77a5c20739a44cc3fbc2"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.233607 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" event={"ID":"5dd65149-6fb3-492b-9719-e5e22e51eedd","Type":"ContainerStarted","Data":"80aaab27054b29cddaf13e567405ab0e9a33ec7949feca31997f81a6da255a12"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.245837 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-kmj2d" podStartSLOduration=194.245797053 podStartE2EDuration="3m14.245797053s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.17494129 +0000 UTC m=+256.487138823" watchObservedRunningTime="2026-03-15 00:11:03.245797053 +0000 UTC m=+256.557994596" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.273505 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.275055 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.775030418 +0000 UTC m=+257.087228141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.275950 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" event={"ID":"1901c2b3-404c-4363-a77d-724f71aedb02","Type":"ContainerStarted","Data":"37fd557c143b1a91fe3b5bc9f54cd72ad2f801311c91637520d2f63fa2baa24c"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.299512 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-zqb9z" podStartSLOduration=7.299488534 podStartE2EDuration="7.299488534s" podCreationTimestamp="2026-03-15 00:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.248075293 +0000 UTC m=+256.560272816" watchObservedRunningTime="2026-03-15 00:11:03.299488534 +0000 UTC m=+256.611686067" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.341319 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" podStartSLOduration=194.341291916 podStartE2EDuration="3m14.341291916s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.298685163 +0000 UTC m=+256.610882696" watchObservedRunningTime="2026-03-15 00:11:03.341291916 +0000 UTC m=+256.653489449" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.356894 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hjq5h" event={"ID":"63356f73-2c6e-472c-a38b-a431c562d4e3","Type":"ContainerStarted","Data":"1c7b5f9a3543dab16838440122fd28cc7ef8d6cd77446b2993309cf0cab5001f"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.356968 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hjq5h" event={"ID":"63356f73-2c6e-472c-a38b-a431c562d4e3","Type":"ContainerStarted","Data":"2754f18e8793c3e976321023acd44aef7ca013da68b7e97a208cdc9374f32f82"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.371056 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" podStartSLOduration=194.371027424 podStartE2EDuration="3m14.371027424s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.369663387 +0000 UTC m=+256.681860920" watchObservedRunningTime="2026-03-15 00:11:03.371027424 +0000 UTC m=+256.683224957" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.374827 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.375256 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.376068 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.376228 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.876204303 +0000 UTC m=+257.188401826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.381920 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.382619 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.882595035 +0000 UTC m=+257.194792568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.426473 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hjq5h" podStartSLOduration=8.426409351 podStartE2EDuration="8.426409351s" podCreationTimestamp="2026-03-15 00:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.413467984 +0000 UTC m=+256.725665517" watchObservedRunningTime="2026-03-15 00:11:03.426409351 +0000 UTC m=+256.738606894" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.467414 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" podStartSLOduration=194.467392211 podStartE2EDuration="3m14.467392211s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.455319477 +0000 UTC m=+256.767517010" watchObservedRunningTime="2026-03-15 00:11:03.467392211 +0000 UTC m=+256.779589744" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.473613 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" event={"ID":"28681c01-d949-429f-ad5c-6084f6d30e81","Type":"ContainerStarted","Data":"476419c8ac13dd1d61e3ee27830eb9e589e42e27f6e896afbddad1fac1e3d1bc"} Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.473738 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7hrpb" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.485922 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk6w5" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.486590 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.487931 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.488399 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:03.988376334 +0000 UTC m=+257.300573867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.489817 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rmkcs" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.491736 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.493972 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.496625 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.526089 4861 ???:1] "http: TLS handshake error from 192.168.126.11:53192: no serving certificate available for the kubelet" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.529772 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.531276 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g2qt2" podStartSLOduration=193.531246405 podStartE2EDuration="3m13.531246405s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.530425723 +0000 UTC m=+256.842623256" watchObservedRunningTime="2026-03-15 00:11:03.531246405 +0000 UTC m=+256.843443938" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.573575 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hrl6d" podStartSLOduration=193.573538251 podStartE2EDuration="3m13.573538251s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.568535896 +0000 UTC m=+256.880733429" watchObservedRunningTime="2026-03-15 00:11:03.573538251 +0000 UTC m=+256.885735784" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.590775 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.590838 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.590864 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdmw\" (UniqueName: \"kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.590924 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.591338 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.091322548 +0000 UTC m=+257.403520081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.599834 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-87hd7" podStartSLOduration=194.599814676 podStartE2EDuration="3m14.599814676s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.593236259 +0000 UTC m=+256.905433792" watchObservedRunningTime="2026-03-15 00:11:03.599814676 +0000 UTC m=+256.912012209" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.665529 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dcj6k" podStartSLOduration=194.66551198 podStartE2EDuration="3m14.66551198s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:03.663182377 +0000 UTC m=+256.975379910" watchObservedRunningTime="2026-03-15 00:11:03.66551198 +0000 UTC m=+256.977709513" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.692688 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.692916 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdmw\" (UniqueName: \"kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.692983 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.693087 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.693703 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.193649075 +0000 UTC m=+257.505846608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.694008 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.694607 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.739308 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.740790 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: W0315 00:11:03.750452 4861 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.750508 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.794485 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.794609 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.794695 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnhzh\" (UniqueName: \"kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.794760 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.795218 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.295204641 +0000 UTC m=+257.607402174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.798710 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:03 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:03 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:03 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.798784 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.812594 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.824064 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdmw\" (UniqueName: \"kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw\") pod \"certified-operators-9rrjp\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.850062 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.855069 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.897135 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.897445 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.897594 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.897650 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnhzh\" (UniqueName: \"kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: E0315 00:11:03.898107 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.398085313 +0000 UTC m=+257.710282856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.898520 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.898794 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.950985 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.952017 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.953091 4861 ???:1] "http: TLS handshake error from 192.168.126.11:43086: no serving certificate available for the kubelet" Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.969937 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:11:03 crc kubenswrapper[4861]: I0315 00:11:03.970587 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnhzh\" (UniqueName: \"kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh\") pod \"community-operators-zz6cr\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.003303 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.003382 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jtc2\" (UniqueName: \"kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.003408 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.003439 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.003784 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.503769069 +0000 UTC m=+257.815966602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.100058 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.101457 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.105190 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.105548 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jtc2\" (UniqueName: \"kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.105603 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.105637 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.106267 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.106369 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.606350593 +0000 UTC m=+257.918548126 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.107156 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.130093 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.208144 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jtc2\" (UniqueName: \"kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2\") pod \"certified-operators-rkmhv\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.209861 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.209907 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpr7r\" (UniqueName: \"kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.209990 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.210010 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.210823 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.71054406 +0000 UTC m=+258.022741593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.311394 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.311730 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpr7r\" (UniqueName: \"kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.311793 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.311811 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.312295 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.312402 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.812374254 +0000 UTC m=+258.124571787 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.313185 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.348960 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.410355 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpr7r\" (UniqueName: \"kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r\") pod \"community-operators-g55tt\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.413817 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.414289 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:04.914275099 +0000 UTC m=+258.226472632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.483208 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-zqb9z" event={"ID":"53dc9c56-f3df-43eb-a8af-f068904d8ab0","Type":"ContainerStarted","Data":"ee05d0f352d93e4c7300628dbffa317539373452044e6c50de20ad07ec2e5a35"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.509701 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" event={"ID":"a2b9027a-9381-417c-b500-16822f802828","Type":"ContainerStarted","Data":"de83e483fd4feddfdfad10a3ee9bf8c290c0bf510605c2d06baa478c079f7c78"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.515082 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.515472 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.015457096 +0000 UTC m=+258.327654629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.526786 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" event={"ID":"7f160fb8-7a29-4f18-a072-ac62270476de","Type":"ContainerStarted","Data":"58fae8146f57b5622ee1edc8a68d8757a942f23537d9ef8eec0d09e12ffff401"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.527864 4861 patch_prober.go:28] interesting pod/console-operator-58897d9998-k9d4t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.527907 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" podUID="7f160fb8-7a29-4f18-a072-ac62270476de" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.566687 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-294fg" podStartSLOduration=195.56666149 podStartE2EDuration="3m15.56666149s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.538953076 +0000 UTC m=+257.851150629" watchObservedRunningTime="2026-03-15 00:11:04.56666149 +0000 UTC m=+257.878859023" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.572168 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" event={"ID":"3589ff40-d99e-4598-8f93-b8721775f117","Type":"ContainerStarted","Data":"76b2b75a4fff83c378d5ae05c80fc476787baebc0e47b4914681c1866031f61e"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.572251 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" event={"ID":"3589ff40-d99e-4598-8f93-b8721775f117","Type":"ContainerStarted","Data":"5a28492eddc4a5fed74e8ce47fe97094766e18a669d987a4120dd4afa031bae3"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.594794 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" event={"ID":"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7","Type":"ContainerStarted","Data":"d674703fb8655d74e628c028f5ba0806897ef01b9296a0d631c751089c1e67b8"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.594853 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" event={"ID":"f4eea3c7-9bac-4ce6-9ae3-e06ab5ffb9f7","Type":"ContainerStarted","Data":"fe9623628677b940a50cf1dd85b1e49f2f93dcb74b6caa36ddcadab351bddc8e"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.595650 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.631150 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.635550 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.135530699 +0000 UTC m=+258.447728242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.637168 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-lns4z" podStartSLOduration=195.637141923 podStartE2EDuration="3m15.637141923s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.635171149 +0000 UTC m=+257.947368692" watchObservedRunningTime="2026-03-15 00:11:04.637141923 +0000 UTC m=+257.949339466" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.638737 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k6lv8" event={"ID":"7d5e0358-4756-4e8c-aec1-baf3f5d52699","Type":"ContainerStarted","Data":"4c775a7d2b94923048ee7ab9b67b6c6352900a90cadf2f3ea958f538fb9cd336"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.638787 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-k6lv8" event={"ID":"7d5e0358-4756-4e8c-aec1-baf3f5d52699","Type":"ContainerStarted","Data":"cc6c5fa20014072e2c1495f446546a6d8705428ddd96897fb913186cb77b8db5"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.639090 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-k6lv8" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.659893 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" event={"ID":"1e8992d1-2a3f-4a3b-b857-821cfb8bed31","Type":"ContainerStarted","Data":"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.660830 4861 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-j2krj container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" start-of-body= Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.660897 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.24:6443/healthz\": dial tcp 10.217.0.24:6443: connect: connection refused" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.682427 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" podStartSLOduration=194.682403718 podStartE2EDuration="3m14.682403718s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.681934505 +0000 UTC m=+257.994132038" watchObservedRunningTime="2026-03-15 00:11:04.682403718 +0000 UTC m=+257.994601241" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.705533 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" event={"ID":"6ef26501-9849-4bd4-8976-38cc771e868e","Type":"ContainerStarted","Data":"085c0578c7aafc340e3717362d03654b5f9897f0bfe123ecc26c78ba18fd6ecb"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.727355 4861 ???:1] "http: TLS handshake error from 192.168.126.11:43098: no serving certificate available for the kubelet" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.733320 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.733952 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.233934364 +0000 UTC m=+258.546131897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.740801 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" event={"ID":"def607d8-9395-4eab-af6b-a03476b186ed","Type":"ContainerStarted","Data":"2335636a7d9e98b444f826f6bf38779e512d60ffea8dd3f1ad0d91a3300106b7"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.740870 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" event={"ID":"def607d8-9395-4eab-af6b-a03476b186ed","Type":"ContainerStarted","Data":"47e917129fde1f05cb54e0e14c3143da67e4f63fe342da1fa6cb5429650701a9"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.743295 4861 generic.go:334] "Generic (PLEG): container finished" podID="59b39d47-b227-436a-9437-688c54bd4e05" containerID="e0e7f47d3c1424256c55a005e859a836b993788ffe7e8e0291df92dd838e856a" exitCode=0 Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.743350 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" event={"ID":"59b39d47-b227-436a-9437-688c54bd4e05","Type":"ContainerDied","Data":"e0e7f47d3c1424256c55a005e859a836b993788ffe7e8e0291df92dd838e856a"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.744473 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" event={"ID":"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1","Type":"ContainerStarted","Data":"6efd979a605f4f4376b4f9b5c86fe6a4d00cd3a2763e7286ba5058ff6435e800"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.745521 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerStarted","Data":"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.747625 4861 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kt6gz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.747696 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.761361 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-k6lv8" podStartSLOduration=8.761339369 podStartE2EDuration="8.761339369s" podCreationTimestamp="2026-03-15 00:10:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.723312527 +0000 UTC m=+258.035510080" watchObservedRunningTime="2026-03-15 00:11:04.761339369 +0000 UTC m=+258.073536902" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.762004 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fbnkg" podStartSLOduration=194.761998528 podStartE2EDuration="3m14.761998528s" podCreationTimestamp="2026-03-15 00:07:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.760734242 +0000 UTC m=+258.072931775" watchObservedRunningTime="2026-03-15 00:11:04.761998528 +0000 UTC m=+258.074196061" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.783794 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" event={"ID":"432faa21-42d4-454a-a36f-dcf427e79e96","Type":"ContainerStarted","Data":"e9f1d1d875919c57fde709c69fe22cfb133d75def422404f7466f5355c88fc11"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.783849 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-7jkmx" event={"ID":"432faa21-42d4-454a-a36f-dcf427e79e96","Type":"ContainerStarted","Data":"4a6f20e7a3d6086604618c05c7e1dc86df71dc55e60b4b38833a9cb243231219"} Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.797152 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:04 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:04 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:04 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.797228 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.813295 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-p295d" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.835984 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.836759 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.840448 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.340427588 +0000 UTC m=+258.652625121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.917000 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5c5ms" podStartSLOduration=195.916977735 podStartE2EDuration="3m15.916977735s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:04.877914845 +0000 UTC m=+258.190112368" watchObservedRunningTime="2026-03-15 00:11:04.916977735 +0000 UTC m=+258.229175268" Mar 15 00:11:04 crc kubenswrapper[4861]: I0315 00:11:04.947878 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:04 crc kubenswrapper[4861]: E0315 00:11:04.950009 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.449989488 +0000 UTC m=+258.762187021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.004627 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.052101 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.063755 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.563716154 +0000 UTC m=+258.875913757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.066025 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.153091 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.153777 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.653752419 +0000 UTC m=+258.965949962 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.154934 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.164084 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.164136 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.256185 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.257078 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.757062594 +0000 UTC m=+259.069260127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.303921 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.364643 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.365020 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.864988528 +0000 UTC m=+259.177186061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.365178 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.365829 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.865822061 +0000 UTC m=+259.178019594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.466702 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.466868 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.966838043 +0000 UTC m=+259.279035586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.467043 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.467509 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:05.967494901 +0000 UTC m=+259.279692434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.576152 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.576916 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.076898107 +0000 UTC m=+259.389095640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.603721 4861 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dh2q9 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.603828 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" podUID="000fdf2c-16b4-4271-a936-3795037adc6e" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.604842 4861 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dh2q9 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.604891 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" podUID="000fdf2c-16b4-4271-a936-3795037adc6e" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.679910 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.680442 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.180424048 +0000 UTC m=+259.492621581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.755536 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.756925 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.761199 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.784037 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.784364 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.28432851 +0000 UTC m=+259.596526043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.784701 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd8xn\" (UniqueName: \"kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.784761 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.784785 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.784858 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.785180 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.285163223 +0000 UTC m=+259.597360756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.792862 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:05 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:05 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:05 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.792941 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.812044 4861 generic.go:334] "Generic (PLEG): container finished" podID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerID="315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f" exitCode=0 Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.812771 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerDied","Data":"315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f"} Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.812852 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerStarted","Data":"74fe1e3996101a89ab628a00d004f4a6b6576500c18cd575b5e58fc70458920d"} Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.839844 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerStarted","Data":"e267de667edd62e9313009121a42e9ddae018f033650fd7f47d13338c1bf4f73"} Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.843021 4861 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kt6gz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.843123 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.887455 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.888433 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.888668 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.388637093 +0000 UTC m=+259.700834626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.889153 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.889361 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd8xn\" (UniqueName: \"kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.889504 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.889623 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.892329 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.392321316 +0000 UTC m=+259.704518849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.899907 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.900072 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.909320 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.946632 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd8xn\" (UniqueName: \"kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn\") pod \"redhat-marketplace-2jfx8\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:05 crc kubenswrapper[4861]: W0315 00:11:05.947548 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3e7027d_525d_48f6_aa18_bb6a622331e6.slice/crio-ff68155b833de13ab267451281245c88c49206e8c1fe79fe87d771773d515c51 WatchSource:0}: Error finding container ff68155b833de13ab267451281245c88c49206e8c1fe79fe87d771773d515c51: Status 404 returned error can't find the container with id ff68155b833de13ab267451281245c88c49206e8c1fe79fe87d771773d515c51 Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.996673 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:05 crc kubenswrapper[4861]: E0315 00:11:05.996937 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.496905487 +0000 UTC m=+259.809103020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:05 crc kubenswrapper[4861]: I0315 00:11:05.997884 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.000838 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.500820207 +0000 UTC m=+259.813017740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.091141 4861 patch_prober.go:28] interesting pod/apiserver-76f77b778f-zw4vw container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]log ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]etcd ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/generic-apiserver-start-informers ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/max-in-flight-filter ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 15 00:11:06 crc kubenswrapper[4861]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [-]poststarthook/project.openshift.io-projectcache failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 15 00:11:06 crc kubenswrapper[4861]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 15 00:11:06 crc kubenswrapper[4861]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 15 00:11:06 crc kubenswrapper[4861]: livez check failed Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.092217 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" podUID="009b6cf3-3e73-4b27-b249-1c72de7bd3dd" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.094701 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.095742 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.100109 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.100445 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.600411888 +0000 UTC m=+259.912609421 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.100861 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.101428 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.601403595 +0000 UTC m=+259.913601138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.103861 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.120601 4861 ???:1] "http: TLS handshake error from 192.168.126.11:43110: no serving certificate available for the kubelet" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.143226 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.202152 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.202312 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.702283582 +0000 UTC m=+260.014481115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.214754 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.215378 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.215871 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqkhg\" (UniqueName: \"kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.216092 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.216872 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.7168503 +0000 UTC m=+260.029047833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.275588 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.320231 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.320956 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.321079 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqkhg\" (UniqueName: \"kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.321125 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.322148 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.322255 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.822232272 +0000 UTC m=+260.134429815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.322505 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.401389 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqkhg\" (UniqueName: \"kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg\") pod \"redhat-marketplace-6hzjv\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.422690 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.423062 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:06.923048178 +0000 UTC m=+260.235245711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.450001 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.527270 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.528398 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.028379639 +0000 UTC m=+260.340577172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.534138 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-k9d4t" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.604972 4861 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-dh2q9 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.605038 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" podUID="000fdf2c-16b4-4271-a936-3795037adc6e" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.632671 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.633061 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.133044933 +0000 UTC m=+260.445242466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.665740 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.683009 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.708766 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.709046 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b39d47-b227-436a-9437-688c54bd4e05" containerName="collect-profiles" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.709061 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b39d47-b227-436a-9437-688c54bd4e05" containerName="collect-profiles" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.714614 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b39d47-b227-436a-9437-688c54bd4e05" containerName="collect-profiles" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.715546 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.721164 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.733966 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") pod \"59b39d47-b227-436a-9437-688c54bd4e05\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734228 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734291 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pvcr\" (UniqueName: \"kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr\") pod \"59b39d47-b227-436a-9437-688c54bd4e05\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734394 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") pod \"59b39d47-b227-436a-9437-688c54bd4e05\" (UID: \"59b39d47-b227-436a-9437-688c54bd4e05\") " Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734683 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9f56\" (UniqueName: \"kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734736 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.734763 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.737325 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.237295075 +0000 UTC m=+260.549492608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.738118 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume" (OuterVolumeSpecName: "config-volume") pod "59b39d47-b227-436a-9437-688c54bd4e05" (UID: "59b39d47-b227-436a-9437-688c54bd4e05"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.762100 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr" (OuterVolumeSpecName: "kube-api-access-5pvcr") pod "59b39d47-b227-436a-9437-688c54bd4e05" (UID: "59b39d47-b227-436a-9437-688c54bd4e05"). InnerVolumeSpecName "kube-api-access-5pvcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.770935 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.773274 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59b39d47-b227-436a-9437-688c54bd4e05" (UID: "59b39d47-b227-436a-9437-688c54bd4e05"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.823632 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:06 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:06 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.823700 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.838763 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9f56\" (UniqueName: \"kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.838817 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.838872 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.838932 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.839024 4861 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59b39d47-b227-436a-9437-688c54bd4e05-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.839036 4861 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59b39d47-b227-436a-9437-688c54bd4e05-config-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.839047 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pvcr\" (UniqueName: \"kubernetes.io/projected/59b39d47-b227-436a-9437-688c54bd4e05-kube-api-access-5pvcr\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.839488 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.840082 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.340069195 +0000 UTC m=+260.652266728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.840316 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.895297 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9f56\" (UniqueName: \"kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56\") pod \"redhat-operators-9447z\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.904498 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerStarted","Data":"16f5f2001d144dfe25262d02ab176d62d97a486d18c9dd42e3f061d3e27819a0"} Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.904589 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerStarted","Data":"be07a8a199b85f0733f8569f594b859d96d7b646af58b128df3cbb5aff553b48"} Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.921421 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerDied","Data":"d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945"} Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.921383 4861 generic.go:334] "Generic (PLEG): container finished" podID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerID="d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945" exitCode=0 Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.922480 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerStarted","Data":"ff68155b833de13ab267451281245c88c49206e8c1fe79fe87d771773d515c51"} Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.940309 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:06 crc kubenswrapper[4861]: E0315 00:11:06.941723 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.441704264 +0000 UTC m=+260.753901797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.941917 4861 generic.go:334] "Generic (PLEG): container finished" podID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerID="bd69696f6a470e446fa7c5784d6076024bd20f5964f89e03951eb305e5a36de1" exitCode=0 Mar 15 00:11:06 crc kubenswrapper[4861]: I0315 00:11:06.943388 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerDied","Data":"bd69696f6a470e446fa7c5784d6076024bd20f5964f89e03951eb305e5a36de1"} Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.038345 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.039611 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" podUID="7447a1d1-672c-4397-99ac-3657c4586eb3" containerName="controller-manager" containerID="cri-o://c1e25c2a27e672103ee399a556a6e78307209eca5395238024e57e6bbd727408" gracePeriod=30 Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.040326 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558880-stl52" event={"ID":"59b39d47-b227-436a-9437-688c54bd4e05","Type":"ContainerDied","Data":"7c815ad3ba32fd9296c648a3c28e56206d31191b6da8d82de0bbf92632ec0652"} Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.040352 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c815ad3ba32fd9296c648a3c28e56206d31191b6da8d82de0bbf92632ec0652" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.040370 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.040478 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerName="route-controller-manager" containerID="cri-o://eb7478619d91d7ecec547ea0be38f3ba57c10422543d246fe82e38ac47837247" gracePeriod=30 Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.044691 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.045435 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.545421671 +0000 UTC m=+260.857619204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.061834 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:11:07 crc kubenswrapper[4861]: W0315 00:11:07.079959 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce66aba6_9ffe_44d9_9c90_961ab7d5a4e1.slice/crio-e77745dd848943e2cf9ae52843c30f6cb6f286373e828d58bf47caf852e5a10c WatchSource:0}: Error finding container e77745dd848943e2cf9ae52843c30f6cb6f286373e828d58bf47caf852e5a10c: Status 404 returned error can't find the container with id e77745dd848943e2cf9ae52843c30f6cb6f286373e828d58bf47caf852e5a10c Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.091841 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.102125 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.102457 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.149411 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.149906 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.649887358 +0000 UTC m=+260.962084891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.252627 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.252661 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.252797 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.252902 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh8k4\" (UniqueName: \"kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.254217 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.754200182 +0000 UTC m=+261.066397715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.348148 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.358184 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.358311 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.858270078 +0000 UTC m=+261.170467611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.362545 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.362622 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.362923 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.363168 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh8k4\" (UniqueName: \"kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.363237 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.863217507 +0000 UTC m=+261.175415040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.443711 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.444846 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.464707 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.465168 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:07.965148743 +0000 UTC m=+261.277346276 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.568271 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.568764 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.068746076 +0000 UTC m=+261.380943609 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.596492 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-dh2q9" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.604315 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.669682 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.669895 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.169858231 +0000 UTC m=+261.482055764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.670535 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.670962 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.17094694 +0000 UTC m=+261.483144473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.695969 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh8k4\" (UniqueName: \"kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4\") pod \"redhat-operators-t4p99\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.771726 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.772809 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.272764044 +0000 UTC m=+261.584961577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.789311 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:07 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:07 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:07 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.789370 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.797678 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.874156 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.874652 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.374631999 +0000 UTC m=+261.686829532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.975164 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.975488 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.475441695 +0000 UTC m=+261.787639228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:07 crc kubenswrapper[4861]: I0315 00:11:07.975782 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:07 crc kubenswrapper[4861]: E0315 00:11:07.976349 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.476337509 +0000 UTC m=+261.788535052 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.072481 4861 generic.go:334] "Generic (PLEG): container finished" podID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerID="eb7478619d91d7ecec547ea0be38f3ba57c10422543d246fe82e38ac47837247" exitCode=0 Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.072549 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" event={"ID":"c6174b71-a5cc-4a0e-902b-9a9091d15d83","Type":"ContainerDied","Data":"eb7478619d91d7ecec547ea0be38f3ba57c10422543d246fe82e38ac47837247"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.074282 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerStarted","Data":"dce5b094404d1b1d833a158d0751742db2e6cb1d05aa5281f15f98cc19af8d1c"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.077377 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.077983 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.577963398 +0000 UTC m=+261.890160931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.078396 4861 generic.go:334] "Generic (PLEG): container finished" podID="dafef788-346c-4a11-a089-accd1a2b088e" containerID="16f5f2001d144dfe25262d02ab176d62d97a486d18c9dd42e3f061d3e27819a0" exitCode=0 Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.078460 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerDied","Data":"16f5f2001d144dfe25262d02ab176d62d97a486d18c9dd42e3f061d3e27819a0"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.088922 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.089944 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerStarted","Data":"e0385562efc1af970c3c699419627f4e0269c93cec174c2b4cd2fd6e0ed5c3fb"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.091191 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerID="6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18" exitCode=0 Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.091241 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerDied","Data":"6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.091256 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerStarted","Data":"e77745dd848943e2cf9ae52843c30f6cb6f286373e828d58bf47caf852e5a10c"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.093250 4861 generic.go:334] "Generic (PLEG): container finished" podID="7447a1d1-672c-4397-99ac-3657c4586eb3" containerID="c1e25c2a27e672103ee399a556a6e78307209eca5395238024e57e6bbd727408" exitCode=0 Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.093274 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" event={"ID":"7447a1d1-672c-4397-99ac-3657c4586eb3","Type":"ContainerDied","Data":"c1e25c2a27e672103ee399a556a6e78307209eca5395238024e57e6bbd727408"} Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.129492 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.179469 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.180971 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.680944034 +0000 UTC m=+261.993141727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.252647 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.253226 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.252722 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.253681 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.283752 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.283825 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zczkp\" (UniqueName: \"kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp\") pod \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.283895 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config\") pod \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.283928 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert\") pod \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.283974 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca\") pod \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\" (UID: \"c6174b71-a5cc-4a0e-902b-9a9091d15d83\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.284988 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.784958839 +0000 UTC m=+262.097156372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.285946 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca" (OuterVolumeSpecName: "client-ca") pod "c6174b71-a5cc-4a0e-902b-9a9091d15d83" (UID: "c6174b71-a5cc-4a0e-902b-9a9091d15d83"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.286102 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config" (OuterVolumeSpecName: "config") pod "c6174b71-a5cc-4a0e-902b-9a9091d15d83" (UID: "c6174b71-a5cc-4a0e-902b-9a9091d15d83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.292263 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp" (OuterVolumeSpecName: "kube-api-access-zczkp") pod "c6174b71-a5cc-4a0e-902b-9a9091d15d83" (UID: "c6174b71-a5cc-4a0e-902b-9a9091d15d83"). InnerVolumeSpecName "kube-api-access-zczkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.292287 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c6174b71-a5cc-4a0e-902b-9a9091d15d83" (UID: "c6174b71-a5cc-4a0e-902b-9a9091d15d83"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.383094 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.385396 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.385512 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.385524 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6174b71-a5cc-4a0e-902b-9a9091d15d83-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.385536 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6174b71-a5cc-4a0e-902b-9a9091d15d83-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.385545 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zczkp\" (UniqueName: \"kubernetes.io/projected/c6174b71-a5cc-4a0e-902b-9a9091d15d83-kube-api-access-zczkp\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.386023 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.885810956 +0000 UTC m=+262.198008489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.391935 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-zw4vw" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.405708 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.486981 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.487194 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.987154506 +0000 UTC m=+262.299352029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.487319 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.488205 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:08.988189625 +0000 UTC m=+262.300387158 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.590834 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config\") pod \"7447a1d1-672c-4397-99ac-3657c4586eb3\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.590965 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2drp\" (UniqueName: \"kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp\") pod \"7447a1d1-672c-4397-99ac-3657c4586eb3\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.591021 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles\") pod \"7447a1d1-672c-4397-99ac-3657c4586eb3\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.591041 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca\") pod \"7447a1d1-672c-4397-99ac-3657c4586eb3\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.591296 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.591380 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert\") pod \"7447a1d1-672c-4397-99ac-3657c4586eb3\" (UID: \"7447a1d1-672c-4397-99ac-3657c4586eb3\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.593213 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.093179107 +0000 UTC m=+262.405376640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.595029 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config" (OuterVolumeSpecName: "config") pod "7447a1d1-672c-4397-99ac-3657c4586eb3" (UID: "7447a1d1-672c-4397-99ac-3657c4586eb3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.595188 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7447a1d1-672c-4397-99ac-3657c4586eb3" (UID: "7447a1d1-672c-4397-99ac-3657c4586eb3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.595543 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca" (OuterVolumeSpecName: "client-ca") pod "7447a1d1-672c-4397-99ac-3657c4586eb3" (UID: "7447a1d1-672c-4397-99ac-3657c4586eb3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.599121 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp" (OuterVolumeSpecName: "kube-api-access-l2drp") pod "7447a1d1-672c-4397-99ac-3657c4586eb3" (UID: "7447a1d1-672c-4397-99ac-3657c4586eb3"). InnerVolumeSpecName "kube-api-access-l2drp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.599698 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7447a1d1-672c-4397-99ac-3657c4586eb3" (UID: "7447a1d1-672c-4397-99ac-3657c4586eb3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693188 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693315 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693334 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2drp\" (UniqueName: \"kubernetes.io/projected/7447a1d1-672c-4397-99ac-3657c4586eb3-kube-api-access-l2drp\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693348 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693362 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7447a1d1-672c-4397-99ac-3657c4586eb3-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.693376 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7447a1d1-672c-4397-99ac-3657c4586eb3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.693659 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.193638153 +0000 UTC m=+262.505835686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.715461 4861 ???:1] "http: TLS handshake error from 192.168.126.11:43120: no serving certificate available for the kubelet" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.782430 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.786376 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:08 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:08 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:08 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.786471 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.794040 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.794697 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.294677466 +0000 UTC m=+262.606874999 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.846981 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.847253 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7447a1d1-672c-4397-99ac-3657c4586eb3" containerName="controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.847272 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="7447a1d1-672c-4397-99ac-3657c4586eb3" containerName="controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.847288 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerName="route-controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.847297 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerName="route-controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.847399 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" containerName="route-controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.847415 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="7447a1d1-672c-4397-99ac-3657c4586eb3" containerName="controller-manager" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.847938 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.849679 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.850185 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.870234 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.895904 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.897702 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.397656451 +0000 UTC m=+262.709853984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.997694 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.997925 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.497887671 +0000 UTC m=+262.810085204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.998078 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.998138 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:08 crc kubenswrapper[4861]: I0315 00:11:08.998195 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:08 crc kubenswrapper[4861]: E0315 00:11:08.998575 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.49853862 +0000 UTC m=+262.810736153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.024126 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.025067 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.025622 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.026101 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.032066 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.034771 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.102775 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.103162 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.103231 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.103809 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.603790329 +0000 UTC m=+262.915987862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.103844 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.120856 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" event={"ID":"c6174b71-a5cc-4a0e-902b-9a9091d15d83","Type":"ContainerDied","Data":"9f7f90b1807b7285d7d41bcca4ae4a01b99a9a05065bce6155ff02068441cec9"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.120931 4861 scope.go:117] "RemoveContainer" containerID="eb7478619d91d7ecec547ea0be38f3ba57c10422543d246fe82e38ac47837247" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.121127 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.129862 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.133397 4861 generic.go:334] "Generic (PLEG): container finished" podID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerID="cb04a90340acb6cf4df240657fb8099c22b2d6589e2d2c3703a0e616c53b36ed" exitCode=0 Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.133476 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerDied","Data":"cb04a90340acb6cf4df240657fb8099c22b2d6589e2d2c3703a0e616c53b36ed"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.138711 4861 generic.go:334] "Generic (PLEG): container finished" podID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerID="83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2" exitCode=0 Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.138757 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerDied","Data":"83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.152768 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.152902 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pnz7" event={"ID":"7447a1d1-672c-4397-99ac-3657c4586eb3","Type":"ContainerDied","Data":"7d271436788ced2b8a789614f8ac4ce249864a7d227b6c0a9bc2e66d4e073217"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.160788 4861 generic.go:334] "Generic (PLEG): container finished" podID="c842b914-5c03-4d7e-85f5-fae121eec542" containerID="fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65" exitCode=0 Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.165589 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerDied","Data":"fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.167969 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.168910 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerStarted","Data":"8fcf7d546d42f7655b97c2d4c7cb8032f11c5a5dad17280b04889c2fda689c7f"} Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.169109 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.182095 4861 scope.go:117] "RemoveContainer" containerID="c1e25c2a27e672103ee399a556a6e78307209eca5395238024e57e6bbd727408" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.182878 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.183065 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.189469 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.207571 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.207640 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjkft\" (UniqueName: \"kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.208919 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.209930 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.210104 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.210411 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.211130 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.211176 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27f8m\" (UniqueName: \"kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.211377 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.211806 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.211962 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.215259 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.715237181 +0000 UTC m=+263.027434714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.316946 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.317304 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.317369 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.817328103 +0000 UTC m=+263.129525636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.317430 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318652 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318717 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjkft\" (UniqueName: \"kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318777 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318828 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318857 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318919 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.318984 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.319012 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.319102 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.319129 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27f8m\" (UniqueName: \"kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.319186 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.321884 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.322472 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.822452286 +0000 UTC m=+263.134649809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.323040 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.323425 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.345317 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.350052 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.381336 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.381550 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27f8m\" (UniqueName: \"kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m\") pod \"controller-manager-ff588666f-9qxt7\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.382137 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjkft\" (UniqueName: \"kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft\") pod \"route-controller-manager-75478584b5-wkkhz\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.383801 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.418274 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.419664 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.419910 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.420025 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.420220 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.420374 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:09.92035386 +0000 UTC m=+263.232551393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.442136 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.450460 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.525438 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pnz7"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.525957 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.525980 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-5fg69"] Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.527515 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.533051 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.533596 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.033578193 +0000 UTC m=+263.345775726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.634105 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.634489 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.1344694 +0000 UTC m=+263.446666933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.735967 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.736461 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.236440558 +0000 UTC m=+263.548638091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.786175 4861 patch_prober.go:28] interesting pod/router-default-5444994796-lgbxx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 15 00:11:09 crc kubenswrapper[4861]: [-]has-synced failed: reason withheld Mar 15 00:11:09 crc kubenswrapper[4861]: [+]process-running ok Mar 15 00:11:09 crc kubenswrapper[4861]: healthz check failed Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.786297 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lgbxx" podUID="631708f5-e5a1-4d6f-8858-60ea4de43cd9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.824059 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:09 crc kubenswrapper[4861]: W0315 00:11:09.836357 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d23ef4c_1d10_4e4d_807b_c6b9febc9fbf.slice/crio-57aacd15e71d2dd1fd0fc892d53cc36f0fca2733f03a13e97b5b0ecb3cd8b364 WatchSource:0}: Error finding container 57aacd15e71d2dd1fd0fc892d53cc36f0fca2733f03a13e97b5b0ecb3cd8b364: Status 404 returned error can't find the container with id 57aacd15e71d2dd1fd0fc892d53cc36f0fca2733f03a13e97b5b0ecb3cd8b364 Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.836724 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.838031 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.337997464 +0000 UTC m=+263.650194997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.838430 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.839485 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.339434224 +0000 UTC m=+263.651631757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.857369 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 15 00:11:09 crc kubenswrapper[4861]: W0315 00:11:09.873003 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfaa90f10_d13f_429c_a79f_c1ed6c325c5b.slice/crio-8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe WatchSource:0}: Error finding container 8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe: Status 404 returned error can't find the container with id 8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.937268 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.937478 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.939211 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:09 crc kubenswrapper[4861]: E0315 00:11:09.939691 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.4395333 +0000 UTC m=+263.751730833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.942087 4861 patch_prober.go:28] interesting pod/console-f9d7485db-4c7wr container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Mar 15 00:11:09 crc kubenswrapper[4861]: I0315 00:11:09.942123 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4c7wr" podUID="b0c9069d-5138-4fd9-9c2f-78b5224287e8" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.044141 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.045060 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.545042747 +0000 UTC m=+263.857240280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.114915 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.122354 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.138354 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.156243 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.156859 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.656832529 +0000 UTC m=+263.969030072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: W0315 00:11:10.166075 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1467c983_00a6_47fe_9fe2_e6edbbea9231.slice/crio-2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89 WatchSource:0}: Error finding container 2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89: Status 404 returned error can't find the container with id 2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89 Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.213385 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"faa90f10-d13f-429c-a79f-c1ed6c325c5b","Type":"ContainerStarted","Data":"8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe"} Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.220347 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" event={"ID":"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf","Type":"ContainerStarted","Data":"57aacd15e71d2dd1fd0fc892d53cc36f0fca2733f03a13e97b5b0ecb3cd8b364"} Mar 15 00:11:10 crc kubenswrapper[4861]: W0315 00:11:10.234833 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacb39e5d_444d_4647_8bc7_e11151d13dd6.slice/crio-7b2d2ad1fce2acc4b4fc489320499780c039900c6daba358dbb7b36d642b9eef WatchSource:0}: Error finding container 7b2d2ad1fce2acc4b4fc489320499780c039900c6daba358dbb7b36d642b9eef: Status 404 returned error can't find the container with id 7b2d2ad1fce2acc4b4fc489320499780c039900c6daba358dbb7b36d642b9eef Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.257830 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.258226 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.75821128 +0000 UTC m=+264.070408813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.375067 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.376515 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.876486934 +0000 UTC m=+264.188684467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.477762 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.478415 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:10.978394939 +0000 UTC m=+264.290592472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.581207 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.581797 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.081776767 +0000 UTC m=+264.393974300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.683203 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.683971 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.18394738 +0000 UTC m=+264.496144913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.784367 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.784922 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.28490328 +0000 UTC m=+264.597100813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.785841 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.789810 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lgbxx" Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.887459 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.887890 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.387873506 +0000 UTC m=+264.700071039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.988458 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.988766 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.488713632 +0000 UTC m=+264.800911165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:10 crc kubenswrapper[4861]: I0315 00:11:10.988989 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:10 crc kubenswrapper[4861]: E0315 00:11:10.990971 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.490943705 +0000 UTC m=+264.803141238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.090054 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.090806 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.590780223 +0000 UTC m=+264.902977756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.192000 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.192482 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.692463052 +0000 UTC m=+265.004660585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.254346 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1467c983-00a6-47fe-9fe2-e6edbbea9231","Type":"ContainerStarted","Data":"42be6350cfb812b418b38c67e0c061bd5deae26223e07715c5f9d29eeab74026"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.254421 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1467c983-00a6-47fe-9fe2-e6edbbea9231","Type":"ContainerStarted","Data":"2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.264639 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" event={"ID":"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1","Type":"ContainerStarted","Data":"4a21b752e8522e822cc424ce2badabdfa19ca651fb631b638a57f0e3028f32dd"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.271798 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"faa90f10-d13f-429c-a79f-c1ed6c325c5b","Type":"ContainerStarted","Data":"39d25bca4301e6f3924ca4bd60634fb0d0b8661ca40b25043c59188e6fb8ee2c"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.300240 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.300643 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" event={"ID":"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf","Type":"ContainerStarted","Data":"fc163f6de5ca1cd47d4a1061e3cc43377d6ab12f29c5f5bec3fcb4f3778c3ab3"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.301037 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.301163 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.801130627 +0000 UTC m=+265.113328150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.333754 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" event={"ID":"acb39e5d-444d-4647-8bc7-e11151d13dd6","Type":"ContainerStarted","Data":"1b9dfb65620e66916da6d1099228dcf4d1b7e6ccacc94d343d8eda418adc124c"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.333854 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" event={"ID":"acb39e5d-444d-4647-8bc7-e11151d13dd6","Type":"ContainerStarted","Data":"7b2d2ad1fce2acc4b4fc489320499780c039900c6daba358dbb7b36d642b9eef"} Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.339881 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" podStartSLOduration=6.339859749 podStartE2EDuration="6.339859749s" podCreationTimestamp="2026-03-15 00:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:11.331172527 +0000 UTC m=+264.643370070" watchObservedRunningTime="2026-03-15 00:11:11.339859749 +0000 UTC m=+264.652057282" Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.405310 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.410091 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:11.91006955 +0000 UTC m=+265.222267083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.414060 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" podStartSLOduration=6.414033571 podStartE2EDuration="6.414033571s" podCreationTimestamp="2026-03-15 00:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:11.393882228 +0000 UTC m=+264.706079761" watchObservedRunningTime="2026-03-15 00:11:11.414033571 +0000 UTC m=+264.726231104" Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.438133 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7447a1d1-672c-4397-99ac-3657c4586eb3" path="/var/lib/kubelet/pods/7447a1d1-672c-4397-99ac-3657c4586eb3/volumes" Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.441049 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6174b71-a5cc-4a0e-902b-9a9091d15d83" path="/var/lib/kubelet/pods/c6174b71-a5cc-4a0e-902b-9a9091d15d83/volumes" Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.441584 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.509752 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.509897 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.009865417 +0000 UTC m=+265.322062950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.510471 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.510924 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.010915307 +0000 UTC m=+265.323112840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.612010 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.612349 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.112316168 +0000 UTC m=+265.424513691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.612409 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.614876 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.11486643 +0000 UTC m=+265.427063963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.713599 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.713954 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.213934587 +0000 UTC m=+265.526132120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.815462 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.818228 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.318212228 +0000 UTC m=+265.630409761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.920308 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:11 crc kubenswrapper[4861]: E0315 00:11:11.921488 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.421450112 +0000 UTC m=+265.733647645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:11 crc kubenswrapper[4861]: I0315 00:11:11.985036 4861 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.027210 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.027733 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.52771338 +0000 UTC m=+265.839910913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.129147 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.129709 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.629691618 +0000 UTC m=+265.941889151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.231888 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.232671 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.732654953 +0000 UTC m=+266.044852486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.236721 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-k6lv8" Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.333162 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.336627 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.836604987 +0000 UTC m=+266.148802520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.447962 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.448667 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:12.948647326 +0000 UTC m=+266.260844859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.498510 4861 generic.go:334] "Generic (PLEG): container finished" podID="1467c983-00a6-47fe-9fe2-e6edbbea9231" containerID="42be6350cfb812b418b38c67e0c061bd5deae26223e07715c5f9d29eeab74026" exitCode=0 Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.498700 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1467c983-00a6-47fe-9fe2-e6edbbea9231","Type":"ContainerDied","Data":"42be6350cfb812b418b38c67e0c061bd5deae26223e07715c5f9d29eeab74026"} Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.538451 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" event={"ID":"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1","Type":"ContainerStarted","Data":"6036592722cafe095400c35122fdfe327a0dd16e0c5b94c4b149221e750886e1"} Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.550543 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.551507 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.051485478 +0000 UTC m=+266.363683011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.555877 4861 generic.go:334] "Generic (PLEG): container finished" podID="faa90f10-d13f-429c-a79f-c1ed6c325c5b" containerID="39d25bca4301e6f3924ca4bd60634fb0d0b8661ca40b25043c59188e6fb8ee2c" exitCode=0 Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.556752 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"faa90f10-d13f-429c-a79f-c1ed6c325c5b","Type":"ContainerDied","Data":"39d25bca4301e6f3924ca4bd60634fb0d0b8661ca40b25043c59188e6fb8ee2c"} Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.557527 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.652691 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.654499 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.154481275 +0000 UTC m=+266.466678808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.679779 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.753892 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.754367 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.254347394 +0000 UTC m=+266.566544927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.856460 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.857004 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.35697138 +0000 UTC m=+266.669168913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.957826 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.957874 4861 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-15T00:11:11.98507164Z","Handler":null,"Name":""} Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.958276 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.458229068 +0000 UTC m=+266.770426601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.958468 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:12 crc kubenswrapper[4861]: E0315 00:11:12.959080 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-15 00:11:13.459069511 +0000 UTC m=+266.771267044 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-7xg99" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.967491 4861 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 15 00:11:12 crc kubenswrapper[4861]: I0315 00:11:12.967830 4861 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.060425 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.064219 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.162713 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.169449 4861 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.169501 4861 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.214063 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-7xg99\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.280312 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.450408 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.865445 4861 ???:1] "http: TLS handshake error from 192.168.126.11:47978: no serving certificate available for the kubelet" Mar 15 00:11:13 crc kubenswrapper[4861]: I0315 00:11:13.968180 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.069974 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.158188 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.213488 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access\") pod \"1467c983-00a6-47fe-9fe2-e6edbbea9231\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.213550 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir\") pod \"1467c983-00a6-47fe-9fe2-e6edbbea9231\" (UID: \"1467c983-00a6-47fe-9fe2-e6edbbea9231\") " Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.213749 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1467c983-00a6-47fe-9fe2-e6edbbea9231" (UID: "1467c983-00a6-47fe-9fe2-e6edbbea9231"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.214041 4861 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1467c983-00a6-47fe-9fe2-e6edbbea9231-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.227237 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1467c983-00a6-47fe-9fe2-e6edbbea9231" (UID: "1467c983-00a6-47fe-9fe2-e6edbbea9231"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.316080 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access\") pod \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.316203 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir\") pod \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\" (UID: \"faa90f10-d13f-429c-a79f-c1ed6c325c5b\") " Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.316360 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "faa90f10-d13f-429c-a79f-c1ed6c325c5b" (UID: "faa90f10-d13f-429c-a79f-c1ed6c325c5b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.316917 4861 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.316932 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1467c983-00a6-47fe-9fe2-e6edbbea9231-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.323056 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "faa90f10-d13f-429c-a79f-c1ed6c325c5b" (UID: "faa90f10-d13f-429c-a79f-c1ed6c325c5b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.419117 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/faa90f10-d13f-429c-a79f-c1ed6c325c5b-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.679798 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"faa90f10-d13f-429c-a79f-c1ed6c325c5b","Type":"ContainerDied","Data":"8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe"} Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.679850 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e21b4130147e5dc911ff9740a1b95b1e7a6286dc45dcae66db2bb77529b5cbe" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.679921 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.706433 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" event={"ID":"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f","Type":"ContainerStarted","Data":"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea"} Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.706482 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" event={"ID":"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f","Type":"ContainerStarted","Data":"50c6d198df9770621577a41f2964020a19e1f2605365ab7a9b9ed45ae8e3427b"} Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.707138 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.724515 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"1467c983-00a6-47fe-9fe2-e6edbbea9231","Type":"ContainerDied","Data":"2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89"} Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.724950 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d8b455b0220d55f567aa87acee481e77f1818e1babf397e7da2bb063d35ae89" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.724779 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.822663 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" event={"ID":"1bfb7c93-e34c-4fe1-a748-371c9bdaa9d1","Type":"ContainerStarted","Data":"c86cf6c5c3c1a3c4d37bd3826ce4d87d7749ce8f1da3125fa4f85af9f398599b"} Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.861266 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" podStartSLOduration=205.861244666 podStartE2EDuration="3m25.861244666s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:14.753920139 +0000 UTC m=+268.066117672" watchObservedRunningTime="2026-03-15 00:11:14.861244666 +0000 UTC m=+268.173442199" Mar 15 00:11:14 crc kubenswrapper[4861]: I0315 00:11:14.861997 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7sdg6" podStartSLOduration=19.861988707 podStartE2EDuration="19.861988707s" podCreationTimestamp="2026-03-15 00:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:14.856887825 +0000 UTC m=+268.169085358" watchObservedRunningTime="2026-03-15 00:11:14.861988707 +0000 UTC m=+268.174186240" Mar 15 00:11:15 crc kubenswrapper[4861]: I0315 00:11:15.635521 4861 ???:1] "http: TLS handshake error from 192.168.126.11:47986: no serving certificate available for the kubelet" Mar 15 00:11:18 crc kubenswrapper[4861]: I0315 00:11:18.252622 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:11:18 crc kubenswrapper[4861]: I0315 00:11:18.253034 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:11:18 crc kubenswrapper[4861]: I0315 00:11:18.252892 4861 patch_prober.go:28] interesting pod/downloads-7954f5f757-xzvqd container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Mar 15 00:11:18 crc kubenswrapper[4861]: I0315 00:11:18.253470 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xzvqd" podUID="2ec532f5-c27d-4a38-900c-fbbc5676a5cd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Mar 15 00:11:19 crc kubenswrapper[4861]: I0315 00:11:19.955677 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:11:19 crc kubenswrapper[4861]: I0315 00:11:19.962240 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4c7wr" Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.110018 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.110967 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerName="controller-manager" containerID="cri-o://1b9dfb65620e66916da6d1099228dcf4d1b7e6ccacc94d343d8eda418adc124c" gracePeriod=30 Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.121409 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.121841 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerName="route-controller-manager" containerID="cri-o://fc163f6de5ca1cd47d4a1061e3cc43377d6ab12f29c5f5bec3fcb4f3778c3ab3" gracePeriod=30 Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.159386 4861 ???:1] "http: TLS handshake error from 192.168.126.11:56634: no serving certificate available for the kubelet" Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.921606 4861 generic.go:334] "Generic (PLEG): container finished" podID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerID="fc163f6de5ca1cd47d4a1061e3cc43377d6ab12f29c5f5bec3fcb4f3778c3ab3" exitCode=0 Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.921854 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" event={"ID":"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf","Type":"ContainerDied","Data":"fc163f6de5ca1cd47d4a1061e3cc43377d6ab12f29c5f5bec3fcb4f3778c3ab3"} Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.925680 4861 generic.go:334] "Generic (PLEG): container finished" podID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerID="1b9dfb65620e66916da6d1099228dcf4d1b7e6ccacc94d343d8eda418adc124c" exitCode=0 Mar 15 00:11:24 crc kubenswrapper[4861]: I0315 00:11:24.925729 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" event={"ID":"acb39e5d-444d-4647-8bc7-e11151d13dd6","Type":"ContainerDied","Data":"1b9dfb65620e66916da6d1099228dcf4d1b7e6ccacc94d343d8eda418adc124c"} Mar 15 00:11:28 crc kubenswrapper[4861]: I0315 00:11:28.262214 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xzvqd" Mar 15 00:11:29 crc kubenswrapper[4861]: I0315 00:11:29.420194 4861 patch_prober.go:28] interesting pod/route-controller-manager-75478584b5-wkkhz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Mar 15 00:11:29 crc kubenswrapper[4861]: I0315 00:11:29.420305 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Mar 15 00:11:29 crc kubenswrapper[4861]: I0315 00:11:29.444505 4861 patch_prober.go:28] interesting pod/controller-manager-ff588666f-9qxt7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Mar 15 00:11:29 crc kubenswrapper[4861]: I0315 00:11:29.444665 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Mar 15 00:11:32 crc kubenswrapper[4861]: I0315 00:11:32.282497 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:11:32 crc kubenswrapper[4861]: I0315 00:11:32.282607 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:11:33 crc kubenswrapper[4861]: I0315 00:11:33.295511 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.946786 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979127 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:35 crc kubenswrapper[4861]: E0315 00:11:35.979461 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa90f10-d13f-429c-a79f-c1ed6c325c5b" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979478 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa90f10-d13f-429c-a79f-c1ed6c325c5b" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: E0315 00:11:35.979497 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1467c983-00a6-47fe-9fe2-e6edbbea9231" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979508 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="1467c983-00a6-47fe-9fe2-e6edbbea9231" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: E0315 00:11:35.979527 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerName="controller-manager" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979541 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerName="controller-manager" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979697 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="1467c983-00a6-47fe-9fe2-e6edbbea9231" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979713 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa90f10-d13f-429c-a79f-c1ed6c325c5b" containerName="pruner" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.979724 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" containerName="controller-manager" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.980290 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:35 crc kubenswrapper[4861]: I0315 00:11:35.986632 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.016479 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" event={"ID":"acb39e5d-444d-4647-8bc7-e11151d13dd6","Type":"ContainerDied","Data":"7b2d2ad1fce2acc4b4fc489320499780c039900c6daba358dbb7b36d642b9eef"} Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.017023 4861 scope.go:117] "RemoveContainer" containerID="1b9dfb65620e66916da6d1099228dcf4d1b7e6ccacc94d343d8eda418adc124c" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.017181 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-ff588666f-9qxt7" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.068154 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert\") pod \"acb39e5d-444d-4647-8bc7-e11151d13dd6\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.068240 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca\") pod \"acb39e5d-444d-4647-8bc7-e11151d13dd6\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.068279 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config\") pod \"acb39e5d-444d-4647-8bc7-e11151d13dd6\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.068338 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles\") pod \"acb39e5d-444d-4647-8bc7-e11151d13dd6\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.068664 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27f8m\" (UniqueName: \"kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m\") pod \"acb39e5d-444d-4647-8bc7-e11151d13dd6\" (UID: \"acb39e5d-444d-4647-8bc7-e11151d13dd6\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069034 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069134 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069188 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069397 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca" (OuterVolumeSpecName: "client-ca") pod "acb39e5d-444d-4647-8bc7-e11151d13dd6" (UID: "acb39e5d-444d-4647-8bc7-e11151d13dd6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069406 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069462 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vqh5\" (UniqueName: \"kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069511 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069524 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "acb39e5d-444d-4647-8bc7-e11151d13dd6" (UID: "acb39e5d-444d-4647-8bc7-e11151d13dd6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.069642 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config" (OuterVolumeSpecName: "config") pod "acb39e5d-444d-4647-8bc7-e11151d13dd6" (UID: "acb39e5d-444d-4647-8bc7-e11151d13dd6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.077548 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m" (OuterVolumeSpecName: "kube-api-access-27f8m") pod "acb39e5d-444d-4647-8bc7-e11151d13dd6" (UID: "acb39e5d-444d-4647-8bc7-e11151d13dd6"). InnerVolumeSpecName "kube-api-access-27f8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.081431 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "acb39e5d-444d-4647-8bc7-e11151d13dd6" (UID: "acb39e5d-444d-4647-8bc7-e11151d13dd6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.170811 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.170873 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vqh5\" (UniqueName: \"kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.170921 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.170942 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.170962 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.171067 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/acb39e5d-444d-4647-8bc7-e11151d13dd6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.171083 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.171093 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/acb39e5d-444d-4647-8bc7-e11151d13dd6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.171106 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27f8m\" (UniqueName: \"kubernetes.io/projected/acb39e5d-444d-4647-8bc7-e11151d13dd6-kube-api-access-27f8m\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.172413 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.172440 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.172522 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.181444 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.188867 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vqh5\" (UniqueName: \"kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5\") pod \"controller-manager-7df8ff5588-dln8t\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.305155 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.359136 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.381760 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-ff588666f-9qxt7"] Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.406502 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.476200 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca\") pod \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.476288 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config\") pod \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.477829 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca" (OuterVolumeSpecName: "client-ca") pod "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" (UID: "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.479077 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config" (OuterVolumeSpecName: "config") pod "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" (UID: "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.577843 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjkft\" (UniqueName: \"kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft\") pod \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.578098 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert\") pod \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\" (UID: \"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf\") " Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.579157 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.579183 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.581331 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft" (OuterVolumeSpecName: "kube-api-access-rjkft") pod "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" (UID: "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf"). InnerVolumeSpecName "kube-api-access-rjkft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.581728 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" (UID: "0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.679801 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:36 crc kubenswrapper[4861]: I0315 00:11:36.679854 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjkft\" (UniqueName: \"kubernetes.io/projected/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf-kube-api-access-rjkft\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.025155 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" event={"ID":"0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf","Type":"ContainerDied","Data":"57aacd15e71d2dd1fd0fc892d53cc36f0fca2733f03a13e97b5b0ecb3cd8b364"} Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.025229 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz" Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.062279 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.065835 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75478584b5-wkkhz"] Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.415890 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" path="/var/lib/kubelet/pods/0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf/volumes" Mar 15 00:11:37 crc kubenswrapper[4861]: I0315 00:11:37.416800 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acb39e5d-444d-4647-8bc7-e11151d13dd6" path="/var/lib/kubelet/pods/acb39e5d-444d-4647-8bc7-e11151d13dd6/volumes" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.034460 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:38 crc kubenswrapper[4861]: E0315 00:11:38.035577 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerName="route-controller-manager" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.035592 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerName="route-controller-manager" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.035733 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d23ef4c-1d10-4e4d-807b-c6b9febc9fbf" containerName="route-controller-manager" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.036313 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.040030 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.040371 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.041265 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.041307 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.041537 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.041691 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.055724 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.202082 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.202139 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.202191 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.202212 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tm7h\" (UniqueName: \"kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.303553 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.303673 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tm7h\" (UniqueName: \"kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.303775 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.303799 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.305806 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.308726 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.323403 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tm7h\" (UniqueName: \"kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.325738 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert\") pod \"route-controller-manager-559b58f7fc-gbcnn\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:38 crc kubenswrapper[4861]: I0315 00:11:38.369293 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:40 crc kubenswrapper[4861]: I0315 00:11:40.135928 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hcv9p" Mar 15 00:11:40 crc kubenswrapper[4861]: E0315 00:11:40.445177 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 15 00:11:40 crc kubenswrapper[4861]: E0315 00:11:40.445751 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4jtc2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rkmhv_openshift-marketplace(c32c6537-08bb-4067-ad4f-7f2ce276be7b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:40 crc kubenswrapper[4861]: E0315 00:11:40.448322 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rkmhv" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" Mar 15 00:11:41 crc kubenswrapper[4861]: E0315 00:11:41.655974 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rkmhv" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" Mar 15 00:11:41 crc kubenswrapper[4861]: E0315 00:11:41.698462 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 15 00:11:41 crc kubenswrapper[4861]: E0315 00:11:41.698673 4861 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 15 00:11:41 crc kubenswrapper[4861]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 15 00:11:41 crc kubenswrapper[4861]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hdf7w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29558890-7spp4_openshift-infra(4585669d-fc84-4c80-9cc3-0185b7fb9ce0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 15 00:11:41 crc kubenswrapper[4861]: > logger="UnhandledError" Mar 15 00:11:41 crc kubenswrapper[4861]: E0315 00:11:41.701452 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29558890-7spp4" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" Mar 15 00:11:42 crc kubenswrapper[4861]: E0315 00:11:42.068986 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29558890-7spp4" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.353842 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.355232 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.361832 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.363900 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.367981 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.389137 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.389219 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.490579 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.490650 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.490744 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.512104 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:43 crc kubenswrapper[4861]: I0315 00:11:43.725084 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:44 crc kubenswrapper[4861]: I0315 00:11:44.110773 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:44 crc kubenswrapper[4861]: I0315 00:11:44.207356 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.596840 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.597065 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.597725 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hnhzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zz6cr_openshift-marketplace(f3e7027d-525d-48f6-aa18-bb6a622331e6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.597779 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zhdmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9rrjp_openshift-marketplace(eaa5b8bf-382f-43a4-a33c-c1002254f2fe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.599087 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9rrjp" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.599090 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zz6cr" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.604029 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.604214 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mh8k4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-t4p99_openshift-marketplace(c842b914-5c03-4d7e-85f5-fae121eec542): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.605664 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-t4p99" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.632935 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.633086 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j9f56,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9447z_openshift-marketplace(0c5fd4bf-9989-488e-83e3-a41e0221c18e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:45 crc kubenswrapper[4861]: E0315 00:11:45.634277 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9447z" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.111093 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9447z" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.111113 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zz6cr" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.111356 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9rrjp" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.111486 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-t4p99" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" Mar 15 00:11:47 crc kubenswrapper[4861]: I0315 00:11:47.119678 4861 scope.go:117] "RemoveContainer" containerID="fc163f6de5ca1cd47d4a1061e3cc43377d6ab12f29c5f5bec3fcb4f3778c3ab3" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.189992 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.190278 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fd8xn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2jfx8_openshift-marketplace(ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.191535 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2jfx8" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.199669 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.199883 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cqkhg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6hzjv_openshift-marketplace(9e466e6e-862b-41f2-9639-fa0adb939f84): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 15 00:11:47 crc kubenswrapper[4861]: E0315 00:11:47.201599 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6hzjv" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" Mar 15 00:11:47 crc kubenswrapper[4861]: I0315 00:11:47.593946 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:47 crc kubenswrapper[4861]: W0315 00:11:47.603646 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67e09e8e_1d48_4b30_955d_5d5f22ebc1d5.slice/crio-4255a1f7d45df5d9b4628a1be19bd7bf54f73af276d072b60b7b2d4e02da913b WatchSource:0}: Error finding container 4255a1f7d45df5d9b4628a1be19bd7bf54f73af276d072b60b7b2d4e02da913b: Status 404 returned error can't find the container with id 4255a1f7d45df5d9b4628a1be19bd7bf54f73af276d072b60b7b2d4e02da913b Mar 15 00:11:47 crc kubenswrapper[4861]: I0315 00:11:47.653175 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:47 crc kubenswrapper[4861]: I0315 00:11:47.655932 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.119111 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" event={"ID":"1f69c9cf-6363-40ab-8533-59cfb2d06a21","Type":"ContainerStarted","Data":"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.119587 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" event={"ID":"1f69c9cf-6363-40ab-8533-59cfb2d06a21","Type":"ContainerStarted","Data":"06b59b48dd7a337ea27c448525c999e53d97fe6225103cff4d6da34dce6f43af"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.119608 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.119260 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" containerID="cri-o://f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979" gracePeriod=30 Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.121307 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" event={"ID":"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5","Type":"ContainerStarted","Data":"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.121359 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" event={"ID":"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5","Type":"ContainerStarted","Data":"4255a1f7d45df5d9b4628a1be19bd7bf54f73af276d072b60b7b2d4e02da913b"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.121367 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" podUID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" containerName="controller-manager" containerID="cri-o://210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029" gracePeriod=30 Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.121472 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.128403 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ebab66f-4173-4d82-a75c-841e8e429a5e","Type":"ContainerStarted","Data":"2f0290aa22921c90e85ec8c4a09e9400b99bfd5ce02443cf7fbaf6996f9d95d3"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.128461 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ebab66f-4173-4d82-a75c-841e8e429a5e","Type":"ContainerStarted","Data":"99fdf76c3cb7fb837b17130b27160d901930bb3f4bc3fb70dee909e49fe3b73c"} Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.133029 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.137350 4861 generic.go:334] "Generic (PLEG): container finished" podID="dafef788-346c-4a11-a089-accd1a2b088e" containerID="da88d1d93266f9da28f93ce1a813e76b2a4d9b7b195c41db99750c5000e76866" exitCode=0 Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.137654 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerDied","Data":"da88d1d93266f9da28f93ce1a813e76b2a4d9b7b195c41db99750c5000e76866"} Mar 15 00:11:48 crc kubenswrapper[4861]: E0315 00:11:48.140137 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6hzjv" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" Mar 15 00:11:48 crc kubenswrapper[4861]: E0315 00:11:48.140137 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2jfx8" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.155071 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" podStartSLOduration=24.155048711 podStartE2EDuration="24.155048711s" podCreationTimestamp="2026-03-15 00:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:48.153674203 +0000 UTC m=+301.465871736" watchObservedRunningTime="2026-03-15 00:11:48.155048711 +0000 UTC m=+301.467246254" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.221587 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=5.221530587 podStartE2EDuration="5.221530587s" podCreationTimestamp="2026-03-15 00:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:48.220851509 +0000 UTC m=+301.533049042" watchObservedRunningTime="2026-03-15 00:11:48.221530587 +0000 UTC m=+301.533728120" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.263054 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" podStartSLOduration=24.263030787 podStartE2EDuration="24.263030787s" podCreationTimestamp="2026-03-15 00:11:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:48.259746864 +0000 UTC m=+301.571944397" watchObservedRunningTime="2026-03-15 00:11:48.263030787 +0000 UTC m=+301.575228310" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.452518 4861 patch_prober.go:28] interesting pod/route-controller-manager-559b58f7fc-gbcnn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": read tcp 10.217.0.2:46230->10.217.0.60:8443: read: connection reset by peer" start-of-body= Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.452904 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": read tcp 10.217.0.2:46230->10.217.0.60:8443: read: connection reset by peer" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.453284 4861 patch_prober.go:28] interesting pod/route-controller-manager-559b58f7fc-gbcnn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" start-of-body= Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.453308 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.60:8443/healthz\": dial tcp 10.217.0.60:8443: connect: connection refused" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.571811 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.624345 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:11:48 crc kubenswrapper[4861]: E0315 00:11:48.624967 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" containerName="controller-manager" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.624991 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" containerName="controller-manager" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.625142 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" containerName="controller-manager" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.626448 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.628192 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.681679 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vqh5\" (UniqueName: \"kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5\") pod \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.681739 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config\") pod \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.681852 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca\") pod \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.681889 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert\") pod \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.681967 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles\") pod \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\" (UID: \"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682107 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d5gm\" (UniqueName: \"kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682155 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682172 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682219 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682260 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.682831 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca" (OuterVolumeSpecName: "client-ca") pod "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" (UID: "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.683082 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config" (OuterVolumeSpecName: "config") pod "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" (UID: "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.683185 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" (UID: "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.743436 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5" (OuterVolumeSpecName: "kube-api-access-2vqh5") pod "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" (UID: "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5"). InnerVolumeSpecName "kube-api-access-2vqh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.743623 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" (UID: "67e09e8e-1d48-4b30-955d-5d5f22ebc1d5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.787359 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d5gm\" (UniqueName: \"kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.787419 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.787460 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.787511 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.787989 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788111 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788126 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788140 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788152 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vqh5\" (UniqueName: \"kubernetes.io/projected/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-kube-api-access-2vqh5\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788164 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788803 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.788935 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.789524 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.794994 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.804072 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d5gm\" (UniqueName: \"kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm\") pod \"controller-manager-6c77bfd598-jmrdm\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.866341 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-559b58f7fc-gbcnn_1f69c9cf-6363-40ab-8533-59cfb2d06a21/route-controller-manager/0.log" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.866453 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.889759 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tm7h\" (UniqueName: \"kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h\") pod \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.889811 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca\") pod \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.889838 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert\") pod \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.889898 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config\") pod \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\" (UID: \"1f69c9cf-6363-40ab-8533-59cfb2d06a21\") " Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.891759 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca" (OuterVolumeSpecName: "client-ca") pod "1f69c9cf-6363-40ab-8533-59cfb2d06a21" (UID: "1f69c9cf-6363-40ab-8533-59cfb2d06a21"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.892567 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config" (OuterVolumeSpecName: "config") pod "1f69c9cf-6363-40ab-8533-59cfb2d06a21" (UID: "1f69c9cf-6363-40ab-8533-59cfb2d06a21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.894346 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h" (OuterVolumeSpecName: "kube-api-access-2tm7h") pod "1f69c9cf-6363-40ab-8533-59cfb2d06a21" (UID: "1f69c9cf-6363-40ab-8533-59cfb2d06a21"). InnerVolumeSpecName "kube-api-access-2tm7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.894973 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1f69c9cf-6363-40ab-8533-59cfb2d06a21" (UID: "1f69c9cf-6363-40ab-8533-59cfb2d06a21"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.952402 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.993761 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.993807 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tm7h\" (UniqueName: \"kubernetes.io/projected/1f69c9cf-6363-40ab-8533-59cfb2d06a21-kube-api-access-2tm7h\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.993821 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1f69c9cf-6363-40ab-8533-59cfb2d06a21-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:48 crc kubenswrapper[4861]: I0315 00:11:48.993838 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f69c9cf-6363-40ab-8533-59cfb2d06a21-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.143134 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146432 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-559b58f7fc-gbcnn_1f69c9cf-6363-40ab-8533-59cfb2d06a21/route-controller-manager/0.log" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146486 4861 generic.go:334] "Generic (PLEG): container finished" podID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerID="f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979" exitCode=255 Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146574 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" event={"ID":"1f69c9cf-6363-40ab-8533-59cfb2d06a21","Type":"ContainerDied","Data":"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146601 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" event={"ID":"1f69c9cf-6363-40ab-8533-59cfb2d06a21","Type":"ContainerDied","Data":"06b59b48dd7a337ea27c448525c999e53d97fe6225103cff4d6da34dce6f43af"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146638 4861 scope.go:117] "RemoveContainer" containerID="f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.146789 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn" Mar 15 00:11:49 crc kubenswrapper[4861]: W0315 00:11:49.151037 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda399224b_3b39_4632_bf97_b85c68ca72db.slice/crio-2c0dc3e5ed00e2956f2dac3503042a833208227e5f362e132bf5ce8c34625717 WatchSource:0}: Error finding container 2c0dc3e5ed00e2956f2dac3503042a833208227e5f362e132bf5ce8c34625717: Status 404 returned error can't find the container with id 2c0dc3e5ed00e2956f2dac3503042a833208227e5f362e132bf5ce8c34625717 Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.153503 4861 generic.go:334] "Generic (PLEG): container finished" podID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" containerID="210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029" exitCode=0 Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.153597 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" event={"ID":"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5","Type":"ContainerDied","Data":"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.153632 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" event={"ID":"67e09e8e-1d48-4b30-955d-5d5f22ebc1d5","Type":"ContainerDied","Data":"4255a1f7d45df5d9b4628a1be19bd7bf54f73af276d072b60b7b2d4e02da913b"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.153680 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7df8ff5588-dln8t" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.164200 4861 generic.go:334] "Generic (PLEG): container finished" podID="9ebab66f-4173-4d82-a75c-841e8e429a5e" containerID="2f0290aa22921c90e85ec8c4a09e9400b99bfd5ce02443cf7fbaf6996f9d95d3" exitCode=0 Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.164287 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ebab66f-4173-4d82-a75c-841e8e429a5e","Type":"ContainerDied","Data":"2f0290aa22921c90e85ec8c4a09e9400b99bfd5ce02443cf7fbaf6996f9d95d3"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.168636 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerStarted","Data":"8654096d2849c07acffd09dc9e60eaa52bc6defa5084dd32bda25ee613c44182"} Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.209815 4861 scope.go:117] "RemoveContainer" containerID="f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979" Mar 15 00:11:49 crc kubenswrapper[4861]: E0315 00:11:49.210383 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979\": container with ID starting with f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979 not found: ID does not exist" containerID="f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.210438 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979"} err="failed to get container status \"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979\": rpc error: code = NotFound desc = could not find container \"f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979\": container with ID starting with f83c578db2d69e58856da5be1f4d278302964380f7ff0b123545c7f7d0dff979 not found: ID does not exist" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.210471 4861 scope.go:117] "RemoveContainer" containerID="210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.231410 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g55tt" podStartSLOduration=4.760311295 podStartE2EDuration="45.231380801s" podCreationTimestamp="2026-03-15 00:11:04 +0000 UTC" firstStartedPulling="2026-03-15 00:11:08.086118376 +0000 UTC m=+261.398315929" lastFinishedPulling="2026-03-15 00:11:48.557187902 +0000 UTC m=+301.869385435" observedRunningTime="2026-03-15 00:11:49.201749063 +0000 UTC m=+302.513946596" watchObservedRunningTime="2026-03-15 00:11:49.231380801 +0000 UTC m=+302.543578354" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.241199 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.242647 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-559b58f7fc-gbcnn"] Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.250365 4861 scope.go:117] "RemoveContainer" containerID="210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029" Mar 15 00:11:49 crc kubenswrapper[4861]: E0315 00:11:49.251222 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029\": container with ID starting with 210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029 not found: ID does not exist" containerID="210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.255897 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029"} err="failed to get container status \"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029\": rpc error: code = NotFound desc = could not find container \"210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029\": container with ID starting with 210443cbd2dd477aa878ba7d274931e4f49ef5116a457ce6b55ed703503c9029 not found: ID does not exist" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.259782 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.264700 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7df8ff5588-dln8t"] Mar 15 00:11:49 crc kubenswrapper[4861]: E0315 00:11:49.314840 4861 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f69c9cf_6363_40ab_8533_59cfb2d06a21.slice/crio-06b59b48dd7a337ea27c448525c999e53d97fe6225103cff4d6da34dce6f43af\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f69c9cf_6363_40ab_8533_59cfb2d06a21.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67e09e8e_1d48_4b30_955d_5d5f22ebc1d5.slice\": RecentStats: unable to find data in memory cache]" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.416516 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" path="/var/lib/kubelet/pods/1f69c9cf-6363-40ab-8533-59cfb2d06a21/volumes" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.417104 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e09e8e-1d48-4b30-955d-5d5f22ebc1d5" path="/var/lib/kubelet/pods/67e09e8e-1d48-4b30-955d-5d5f22ebc1d5/volumes" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.545377 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 15 00:11:49 crc kubenswrapper[4861]: E0315 00:11:49.545739 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.545767 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.545926 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f69c9cf-6363-40ab-8533-59cfb2d06a21" containerName="route-controller-manager" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.546440 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.563803 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.602202 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.602267 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.602459 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.703774 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.704252 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.704325 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.704366 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.704484 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.724406 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access\") pod \"installer-9-crc\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:49 crc kubenswrapper[4861]: I0315 00:11:49.895010 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.142077 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.183442 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7d821af9-853a-43b8-8a31-c7df6895c19a","Type":"ContainerStarted","Data":"173f6b46cba544afa2e76cbbe0dd44c46a7f616b5e7ed8fff497e7ea3a8251d2"} Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.190313 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" event={"ID":"a399224b-3b39-4632-bf97-b85c68ca72db","Type":"ContainerStarted","Data":"8b1867ceafb195bd87c05c0528678993af7629ed100bb89d945155420e32160d"} Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.190380 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" event={"ID":"a399224b-3b39-4632-bf97-b85c68ca72db","Type":"ContainerStarted","Data":"2c0dc3e5ed00e2956f2dac3503042a833208227e5f362e132bf5ce8c34625717"} Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.506251 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.522626 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" podStartSLOduration=6.522603802 podStartE2EDuration="6.522603802s" podCreationTimestamp="2026-03-15 00:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:50.207816202 +0000 UTC m=+303.520013735" watchObservedRunningTime="2026-03-15 00:11:50.522603802 +0000 UTC m=+303.834801335" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.622046 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir\") pod \"9ebab66f-4173-4d82-a75c-841e8e429a5e\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.622193 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ebab66f-4173-4d82-a75c-841e8e429a5e" (UID: "9ebab66f-4173-4d82-a75c-841e8e429a5e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.622219 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access\") pod \"9ebab66f-4173-4d82-a75c-841e8e429a5e\" (UID: \"9ebab66f-4173-4d82-a75c-841e8e429a5e\") " Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.623049 4861 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ebab66f-4173-4d82-a75c-841e8e429a5e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.630425 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ebab66f-4173-4d82-a75c-841e8e429a5e" (UID: "9ebab66f-4173-4d82-a75c-841e8e429a5e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:50 crc kubenswrapper[4861]: I0315 00:11:50.726272 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ebab66f-4173-4d82-a75c-841e8e429a5e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.044196 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:11:51 crc kubenswrapper[4861]: E0315 00:11:51.044465 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ebab66f-4173-4d82-a75c-841e8e429a5e" containerName="pruner" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.044479 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ebab66f-4173-4d82-a75c-841e8e429a5e" containerName="pruner" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.044621 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ebab66f-4173-4d82-a75c-841e8e429a5e" containerName="pruner" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.045104 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.046824 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.047365 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.047811 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.048301 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.048684 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.049504 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.058876 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.135800 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47trg\" (UniqueName: \"kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.135879 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.135960 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.136019 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.197620 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7d821af9-853a-43b8-8a31-c7df6895c19a","Type":"ContainerStarted","Data":"13e60e578154bdfac6c323436f2ebd375b86188e7a57c4fb1cdba0c4c1c755bd"} Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.200106 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"9ebab66f-4173-4d82-a75c-841e8e429a5e","Type":"ContainerDied","Data":"99fdf76c3cb7fb837b17130b27160d901930bb3f4bc3fb70dee909e49fe3b73c"} Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.200186 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99fdf76c3cb7fb837b17130b27160d901930bb3f4bc3fb70dee909e49fe3b73c" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.200439 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.200736 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.206507 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.237239 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.237441 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.237497 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47trg\" (UniqueName: \"kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.237581 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.238618 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.239443 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.241968 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.252872 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.252832907 podStartE2EDuration="2.252832907s" podCreationTimestamp="2026-03-15 00:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:51.220078443 +0000 UTC m=+304.532275986" watchObservedRunningTime="2026-03-15 00:11:51.252832907 +0000 UTC m=+304.565030440" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.263364 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47trg\" (UniqueName: \"kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg\") pod \"route-controller-manager-688bf79b7d-ztr76\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.399948 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:51 crc kubenswrapper[4861]: I0315 00:11:51.814254 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:11:51 crc kubenswrapper[4861]: W0315 00:11:51.822775 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57dec95d_ab68_4794_9644_96e48cf6016c.slice/crio-cec2e29d36524ef22c8686e00f96402914b7d7eed2ce879c1042458d585b0e48 WatchSource:0}: Error finding container cec2e29d36524ef22c8686e00f96402914b7d7eed2ce879c1042458d585b0e48: Status 404 returned error can't find the container with id cec2e29d36524ef22c8686e00f96402914b7d7eed2ce879c1042458d585b0e48 Mar 15 00:11:52 crc kubenswrapper[4861]: I0315 00:11:52.220164 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" event={"ID":"57dec95d-ab68-4794-9644-96e48cf6016c","Type":"ContainerStarted","Data":"546baeb43860ca258b27ee387fbc56c722b805d95fbc43b59c6ecad6128a5e3e"} Mar 15 00:11:52 crc kubenswrapper[4861]: I0315 00:11:52.220243 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" event={"ID":"57dec95d-ab68-4794-9644-96e48cf6016c","Type":"ContainerStarted","Data":"cec2e29d36524ef22c8686e00f96402914b7d7eed2ce879c1042458d585b0e48"} Mar 15 00:11:52 crc kubenswrapper[4861]: I0315 00:11:52.238750 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" podStartSLOduration=8.238728631 podStartE2EDuration="8.238728631s" podCreationTimestamp="2026-03-15 00:11:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:11:52.237356223 +0000 UTC m=+305.549553756" watchObservedRunningTime="2026-03-15 00:11:52.238728631 +0000 UTC m=+305.550926164" Mar 15 00:11:53 crc kubenswrapper[4861]: I0315 00:11:53.225882 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:53 crc kubenswrapper[4861]: I0315 00:11:53.232984 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:11:55 crc kubenswrapper[4861]: I0315 00:11:55.164736 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:55 crc kubenswrapper[4861]: I0315 00:11:55.165259 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:55 crc kubenswrapper[4861]: I0315 00:11:55.320192 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:55 crc kubenswrapper[4861]: I0315 00:11:55.361871 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:55 crc kubenswrapper[4861]: I0315 00:11:55.553077 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:57 crc kubenswrapper[4861]: I0315 00:11:57.251152 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g55tt" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="registry-server" containerID="cri-o://8654096d2849c07acffd09dc9e60eaa52bc6defa5084dd32bda25ee613c44182" gracePeriod=2 Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.261639 4861 generic.go:334] "Generic (PLEG): container finished" podID="dafef788-346c-4a11-a089-accd1a2b088e" containerID="8654096d2849c07acffd09dc9e60eaa52bc6defa5084dd32bda25ee613c44182" exitCode=0 Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.261691 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerDied","Data":"8654096d2849c07acffd09dc9e60eaa52bc6defa5084dd32bda25ee613c44182"} Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.647642 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.787887 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpr7r\" (UniqueName: \"kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r\") pod \"dafef788-346c-4a11-a089-accd1a2b088e\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.788051 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities\") pod \"dafef788-346c-4a11-a089-accd1a2b088e\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.788186 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content\") pod \"dafef788-346c-4a11-a089-accd1a2b088e\" (UID: \"dafef788-346c-4a11-a089-accd1a2b088e\") " Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.789433 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities" (OuterVolumeSpecName: "utilities") pod "dafef788-346c-4a11-a089-accd1a2b088e" (UID: "dafef788-346c-4a11-a089-accd1a2b088e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.798694 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r" (OuterVolumeSpecName: "kube-api-access-jpr7r") pod "dafef788-346c-4a11-a089-accd1a2b088e" (UID: "dafef788-346c-4a11-a089-accd1a2b088e"). InnerVolumeSpecName "kube-api-access-jpr7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.890871 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpr7r\" (UniqueName: \"kubernetes.io/projected/dafef788-346c-4a11-a089-accd1a2b088e-kube-api-access-jpr7r\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.890925 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.977840 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dafef788-346c-4a11-a089-accd1a2b088e" (UID: "dafef788-346c-4a11-a089-accd1a2b088e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:11:58 crc kubenswrapper[4861]: I0315 00:11:58.992505 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dafef788-346c-4a11-a089-accd1a2b088e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.271872 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558890-7spp4" event={"ID":"4585669d-fc84-4c80-9cc3-0185b7fb9ce0","Type":"ContainerStarted","Data":"b303a70c6a8191095300a76be411edd13d0c8cbf766fd935a1db3236689f6d3e"} Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.275091 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerStarted","Data":"835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6"} Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.278687 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g55tt" event={"ID":"dafef788-346c-4a11-a089-accd1a2b088e","Type":"ContainerDied","Data":"be07a8a199b85f0733f8569f594b859d96d7b646af58b128df3cbb5aff553b48"} Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.278761 4861 scope.go:117] "RemoveContainer" containerID="8654096d2849c07acffd09dc9e60eaa52bc6defa5084dd32bda25ee613c44182" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.278898 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g55tt" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.304298 4861 scope.go:117] "RemoveContainer" containerID="da88d1d93266f9da28f93ce1a813e76b2a4d9b7b195c41db99750c5000e76866" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.329217 4861 scope.go:117] "RemoveContainer" containerID="16f5f2001d144dfe25262d02ab176d62d97a486d18c9dd42e3f061d3e27819a0" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.333118 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.336817 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g55tt"] Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.416873 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dafef788-346c-4a11-a089-accd1a2b088e" path="/var/lib/kubelet/pods/dafef788-346c-4a11-a089-accd1a2b088e/volumes" Mar 15 00:11:59 crc kubenswrapper[4861]: E0315 00:11:59.451649 4861 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32c6537_08bb_4067_ad4f_7f2ce276be7b.slice/crio-835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32c6537_08bb_4067_ad4f_7f2ce276be7b.slice/crio-conmon-835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6.scope\": RecentStats: unable to find data in memory cache]" Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.737468 4861 csr.go:261] certificate signing request csr-dlmtd is approved, waiting to be issued Mar 15 00:11:59 crc kubenswrapper[4861]: I0315 00:11:59.743998 4861 csr.go:257] certificate signing request csr-dlmtd is issued Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.147138 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558892-8vbws"] Mar 15 00:12:00 crc kubenswrapper[4861]: E0315 00:12:00.147407 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="extract-utilities" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.147421 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="extract-utilities" Mar 15 00:12:00 crc kubenswrapper[4861]: E0315 00:12:00.147437 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="registry-server" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.147443 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="registry-server" Mar 15 00:12:00 crc kubenswrapper[4861]: E0315 00:12:00.147449 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="extract-content" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.147455 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="extract-content" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.147758 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafef788-346c-4a11-a089-accd1a2b088e" containerName="registry-server" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.148187 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.152093 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.191387 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558892-8vbws"] Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.213826 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5jpk\" (UniqueName: \"kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk\") pod \"auto-csr-approver-29558892-8vbws\" (UID: \"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5\") " pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.285833 4861 generic.go:334] "Generic (PLEG): container finished" podID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" containerID="b303a70c6a8191095300a76be411edd13d0c8cbf766fd935a1db3236689f6d3e" exitCode=0 Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.286616 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558890-7spp4" event={"ID":"4585669d-fc84-4c80-9cc3-0185b7fb9ce0","Type":"ContainerDied","Data":"b303a70c6a8191095300a76be411edd13d0c8cbf766fd935a1db3236689f6d3e"} Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.287870 4861 generic.go:334] "Generic (PLEG): container finished" podID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerID="835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6" exitCode=0 Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.287916 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerDied","Data":"835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6"} Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.314970 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5jpk\" (UniqueName: \"kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk\") pod \"auto-csr-approver-29558892-8vbws\" (UID: \"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5\") " pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.345398 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5jpk\" (UniqueName: \"kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk\") pod \"auto-csr-approver-29558892-8vbws\" (UID: \"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5\") " pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.462103 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.745848 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-21 19:41:57.687440958 +0000 UTC Mar 15 00:12:00 crc kubenswrapper[4861]: I0315 00:12:00.746368 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6043h29m56.941077382s for next certificate rotation Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.297568 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558892-8vbws"] Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.301419 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerStarted","Data":"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998"} Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.310568 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerStarted","Data":"01cd5e0d856603c45a8198ddaccfe97a9b6201cf9977c99be55d5a512645baf0"} Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.349056 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rkmhv" podStartSLOduration=4.338995891 podStartE2EDuration="58.349034979s" podCreationTimestamp="2026-03-15 00:11:03 +0000 UTC" firstStartedPulling="2026-03-15 00:11:06.951862218 +0000 UTC m=+260.264059751" lastFinishedPulling="2026-03-15 00:12:00.961901306 +0000 UTC m=+314.274098839" observedRunningTime="2026-03-15 00:12:01.345679634 +0000 UTC m=+314.657877177" watchObservedRunningTime="2026-03-15 00:12:01.349034979 +0000 UTC m=+314.661232502" Mar 15 00:12:01 crc kubenswrapper[4861]: W0315 00:12:01.375869 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9efc1aaf_9347_4ed0_96ea_efa5342aa0b5.slice/crio-d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a WatchSource:0}: Error finding container d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a: Status 404 returned error can't find the container with id d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.718625 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.747248 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-13 19:32:22.577127527 +0000 UTC Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.747302 4861 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6571h20m20.829829416s for next certificate rotation Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.852792 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdf7w\" (UniqueName: \"kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w\") pod \"4585669d-fc84-4c80-9cc3-0185b7fb9ce0\" (UID: \"4585669d-fc84-4c80-9cc3-0185b7fb9ce0\") " Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.860312 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w" (OuterVolumeSpecName: "kube-api-access-hdf7w") pod "4585669d-fc84-4c80-9cc3-0185b7fb9ce0" (UID: "4585669d-fc84-4c80-9cc3-0185b7fb9ce0"). InnerVolumeSpecName "kube-api-access-hdf7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:01 crc kubenswrapper[4861]: I0315 00:12:01.955246 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdf7w\" (UniqueName: \"kubernetes.io/projected/4585669d-fc84-4c80-9cc3-0185b7fb9ce0-kube-api-access-hdf7w\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.281978 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.282038 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.282095 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.282739 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.282800 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390" gracePeriod=600 Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.320988 4861 generic.go:334] "Generic (PLEG): container finished" podID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerID="b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998" exitCode=0 Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.321075 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerDied","Data":"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.324090 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerStarted","Data":"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.330295 4861 generic.go:334] "Generic (PLEG): container finished" podID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerID="6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305" exitCode=0 Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.330359 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerDied","Data":"6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.334731 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerStarted","Data":"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.342094 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerID="f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98" exitCode=0 Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.342203 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerDied","Data":"f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.354323 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558890-7spp4" event={"ID":"4585669d-fc84-4c80-9cc3-0185b7fb9ce0","Type":"ContainerDied","Data":"c63ca3f0ec4db37a08dae1b2ac77bfdb8b254e622c0e466843fef8aa3c420313"} Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.354367 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558890-7spp4" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.354381 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c63ca3f0ec4db37a08dae1b2ac77bfdb8b254e622c0e466843fef8aa3c420313" Mar 15 00:12:02 crc kubenswrapper[4861]: I0315 00:12:02.356664 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558892-8vbws" event={"ID":"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5","Type":"ContainerStarted","Data":"d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.364889 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390" exitCode=0 Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.365084 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.365936 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.369915 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerStarted","Data":"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.372597 4861 generic.go:334] "Generic (PLEG): container finished" podID="c842b914-5c03-4d7e-85f5-fae121eec542" containerID="d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43" exitCode=0 Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.372660 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerDied","Data":"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.374781 4861 generic.go:334] "Generic (PLEG): container finished" podID="9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" containerID="43783405a59596ff4b17a60a37253b280aa04a127834d2ce8745dcc3108822c9" exitCode=0 Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.374946 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558892-8vbws" event={"ID":"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5","Type":"ContainerDied","Data":"43783405a59596ff4b17a60a37253b280aa04a127834d2ce8745dcc3108822c9"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.378659 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerStarted","Data":"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.381641 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerStarted","Data":"21235c8b1e36b19a097af2b256db464f1d71853780199e8fcdbf8d5f87cc9c8b"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.384966 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerStarted","Data":"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.387382 4861 generic.go:334] "Generic (PLEG): container finished" podID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerID="74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719" exitCode=0 Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.387472 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerDied","Data":"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719"} Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.425335 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9rrjp" podStartSLOduration=3.418614785 podStartE2EDuration="1m0.425314106s" podCreationTimestamp="2026-03-15 00:11:03 +0000 UTC" firstStartedPulling="2026-03-15 00:11:05.826175319 +0000 UTC m=+259.138372842" lastFinishedPulling="2026-03-15 00:12:02.83287463 +0000 UTC m=+316.145072163" observedRunningTime="2026-03-15 00:12:03.420365997 +0000 UTC m=+316.732563530" watchObservedRunningTime="2026-03-15 00:12:03.425314106 +0000 UTC m=+316.737511639" Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.467172 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2jfx8" podStartSLOduration=3.780611371 podStartE2EDuration="58.467143544s" podCreationTimestamp="2026-03-15 00:11:05 +0000 UTC" firstStartedPulling="2026-03-15 00:11:08.093738108 +0000 UTC m=+261.405935641" lastFinishedPulling="2026-03-15 00:12:02.780270281 +0000 UTC m=+316.092467814" observedRunningTime="2026-03-15 00:12:03.450713335 +0000 UTC m=+316.762910878" watchObservedRunningTime="2026-03-15 00:12:03.467143544 +0000 UTC m=+316.779341077" Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.562597 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zz6cr" podStartSLOduration=4.777182937 podStartE2EDuration="1m0.562536798s" podCreationTimestamp="2026-03-15 00:11:03 +0000 UTC" firstStartedPulling="2026-03-15 00:11:06.933236887 +0000 UTC m=+260.245434420" lastFinishedPulling="2026-03-15 00:12:02.718590738 +0000 UTC m=+316.030788281" observedRunningTime="2026-03-15 00:12:03.559369939 +0000 UTC m=+316.871567482" watchObservedRunningTime="2026-03-15 00:12:03.562536798 +0000 UTC m=+316.874734331" Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.851366 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:12:03 crc kubenswrapper[4861]: I0315 00:12:03.851865 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.154734 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.155036 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" podUID="a399224b-3b39-4632-bf97-b85c68ca72db" containerName="controller-manager" containerID="cri-o://8b1867ceafb195bd87c05c0528678993af7629ed100bb89d945155420e32160d" gracePeriod=30 Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.190578 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.190827 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" podUID="57dec95d-ab68-4794-9644-96e48cf6016c" containerName="route-controller-manager" containerID="cri-o://546baeb43860ca258b27ee387fbc56c722b805d95fbc43b59c6ecad6128a5e3e" gracePeriod=30 Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.352958 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.353574 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.398318 4861 generic.go:334] "Generic (PLEG): container finished" podID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerID="21235c8b1e36b19a097af2b256db464f1d71853780199e8fcdbf8d5f87cc9c8b" exitCode=0 Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.398375 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerDied","Data":"21235c8b1e36b19a097af2b256db464f1d71853780199e8fcdbf8d5f87cc9c8b"} Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.417129 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" event={"ID":"a399224b-3b39-4632-bf97-b85c68ca72db","Type":"ContainerDied","Data":"8b1867ceafb195bd87c05c0528678993af7629ed100bb89d945155420e32160d"} Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.417216 4861 generic.go:334] "Generic (PLEG): container finished" podID="a399224b-3b39-4632-bf97-b85c68ca72db" containerID="8b1867ceafb195bd87c05c0528678993af7629ed100bb89d945155420e32160d" exitCode=0 Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.428275 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerStarted","Data":"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc"} Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.430717 4861 generic.go:334] "Generic (PLEG): container finished" podID="57dec95d-ab68-4794-9644-96e48cf6016c" containerID="546baeb43860ca258b27ee387fbc56c722b805d95fbc43b59c6ecad6128a5e3e" exitCode=0 Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.430795 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" event={"ID":"57dec95d-ab68-4794-9644-96e48cf6016c","Type":"ContainerDied","Data":"546baeb43860ca258b27ee387fbc56c722b805d95fbc43b59c6ecad6128a5e3e"} Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.431977 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.433745 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerStarted","Data":"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847"} Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.455640 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6hzjv" podStartSLOduration=3.830956148 podStartE2EDuration="58.455614171s" podCreationTimestamp="2026-03-15 00:11:06 +0000 UTC" firstStartedPulling="2026-03-15 00:11:09.14070162 +0000 UTC m=+262.452899143" lastFinishedPulling="2026-03-15 00:12:03.765359633 +0000 UTC m=+317.077557166" observedRunningTime="2026-03-15 00:12:04.45346347 +0000 UTC m=+317.765661003" watchObservedRunningTime="2026-03-15 00:12:04.455614171 +0000 UTC m=+317.767811704" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.497666 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t4p99" podStartSLOduration=2.8387126350000003 podStartE2EDuration="57.497643815s" podCreationTimestamp="2026-03-15 00:11:07 +0000 UTC" firstStartedPulling="2026-03-15 00:11:09.182567449 +0000 UTC m=+262.494764982" lastFinishedPulling="2026-03-15 00:12:03.841498629 +0000 UTC m=+317.153696162" observedRunningTime="2026-03-15 00:12:04.497087188 +0000 UTC m=+317.809284721" watchObservedRunningTime="2026-03-15 00:12:04.497643815 +0000 UTC m=+317.809841338" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.815330 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.892104 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.902306 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47trg\" (UniqueName: \"kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg\") pod \"57dec95d-ab68-4794-9644-96e48cf6016c\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.902430 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert\") pod \"57dec95d-ab68-4794-9644-96e48cf6016c\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.902498 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config\") pod \"57dec95d-ab68-4794-9644-96e48cf6016c\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.902550 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca\") pod \"57dec95d-ab68-4794-9644-96e48cf6016c\" (UID: \"57dec95d-ab68-4794-9644-96e48cf6016c\") " Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.903427 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca" (OuterVolumeSpecName: "client-ca") pod "57dec95d-ab68-4794-9644-96e48cf6016c" (UID: "57dec95d-ab68-4794-9644-96e48cf6016c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.903471 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config" (OuterVolumeSpecName: "config") pod "57dec95d-ab68-4794-9644-96e48cf6016c" (UID: "57dec95d-ab68-4794-9644-96e48cf6016c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.913511 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "57dec95d-ab68-4794-9644-96e48cf6016c" (UID: "57dec95d-ab68-4794-9644-96e48cf6016c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.914791 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg" (OuterVolumeSpecName: "kube-api-access-47trg") pod "57dec95d-ab68-4794-9644-96e48cf6016c" (UID: "57dec95d-ab68-4794-9644-96e48cf6016c"). InnerVolumeSpecName "kube-api-access-47trg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.930141 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9rrjp" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="registry-server" probeResult="failure" output=< Mar 15 00:12:04 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:12:04 crc kubenswrapper[4861]: > Mar 15 00:12:04 crc kubenswrapper[4861]: I0315 00:12:04.964192 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004152 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca\") pod \"a399224b-3b39-4632-bf97-b85c68ca72db\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004285 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d5gm\" (UniqueName: \"kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm\") pod \"a399224b-3b39-4632-bf97-b85c68ca72db\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004337 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config\") pod \"a399224b-3b39-4632-bf97-b85c68ca72db\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004362 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles\") pod \"a399224b-3b39-4632-bf97-b85c68ca72db\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004477 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert\") pod \"a399224b-3b39-4632-bf97-b85c68ca72db\" (UID: \"a399224b-3b39-4632-bf97-b85c68ca72db\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004818 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57dec95d-ab68-4794-9644-96e48cf6016c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004832 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004845 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57dec95d-ab68-4794-9644-96e48cf6016c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.004856 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47trg\" (UniqueName: \"kubernetes.io/projected/57dec95d-ab68-4794-9644-96e48cf6016c-kube-api-access-47trg\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.006088 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config" (OuterVolumeSpecName: "config") pod "a399224b-3b39-4632-bf97-b85c68ca72db" (UID: "a399224b-3b39-4632-bf97-b85c68ca72db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.006367 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca" (OuterVolumeSpecName: "client-ca") pod "a399224b-3b39-4632-bf97-b85c68ca72db" (UID: "a399224b-3b39-4632-bf97-b85c68ca72db"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.006905 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a399224b-3b39-4632-bf97-b85c68ca72db" (UID: "a399224b-3b39-4632-bf97-b85c68ca72db"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.010660 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a399224b-3b39-4632-bf97-b85c68ca72db" (UID: "a399224b-3b39-4632-bf97-b85c68ca72db"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.012735 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm" (OuterVolumeSpecName: "kube-api-access-5d5gm") pod "a399224b-3b39-4632-bf97-b85c68ca72db" (UID: "a399224b-3b39-4632-bf97-b85c68ca72db"). InnerVolumeSpecName "kube-api-access-5d5gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.105625 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5jpk\" (UniqueName: \"kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk\") pod \"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5\" (UID: \"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5\") " Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.106085 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d5gm\" (UniqueName: \"kubernetes.io/projected/a399224b-3b39-4632-bf97-b85c68ca72db-kube-api-access-5d5gm\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.106105 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.106118 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.106129 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a399224b-3b39-4632-bf97-b85c68ca72db-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.106138 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a399224b-3b39-4632-bf97-b85c68ca72db-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.112681 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk" (OuterVolumeSpecName: "kube-api-access-x5jpk") pod "9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" (UID: "9efc1aaf-9347-4ed0-96ea-efa5342aa0b5"). InnerVolumeSpecName "kube-api-access-x5jpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.165640 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.165715 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.207626 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5jpk\" (UniqueName: \"kubernetes.io/projected/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5-kube-api-access-x5jpk\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.446491 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558892-8vbws" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.446511 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558892-8vbws" event={"ID":"9efc1aaf-9347-4ed0-96ea-efa5342aa0b5","Type":"ContainerDied","Data":"d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a"} Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.447429 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d41693957483edcaf914ded124b936d7374976cee8318ee0057c8db108b90b7a" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.449476 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerStarted","Data":"69e3a76c57af1a427485c1b19a18e50fc576f41ba430b2cb8cc0b770cfea5966"} Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.451364 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" event={"ID":"a399224b-3b39-4632-bf97-b85c68ca72db","Type":"ContainerDied","Data":"2c0dc3e5ed00e2956f2dac3503042a833208227e5f362e132bf5ce8c34625717"} Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.451402 4861 scope.go:117] "RemoveContainer" containerID="8b1867ceafb195bd87c05c0528678993af7629ed100bb89d945155420e32160d" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.451499 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c77bfd598-jmrdm" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.456023 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" event={"ID":"57dec95d-ab68-4794-9644-96e48cf6016c","Type":"ContainerDied","Data":"cec2e29d36524ef22c8686e00f96402914b7d7eed2ce879c1042458d585b0e48"} Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.456132 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.470914 4861 scope.go:117] "RemoveContainer" containerID="546baeb43860ca258b27ee387fbc56c722b805d95fbc43b59c6ecad6128a5e3e" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.475892 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9447z" podStartSLOduration=3.7094286419999998 podStartE2EDuration="59.475854124s" podCreationTimestamp="2026-03-15 00:11:06 +0000 UTC" firstStartedPulling="2026-03-15 00:11:09.137110649 +0000 UTC m=+262.449308182" lastFinishedPulling="2026-03-15 00:12:04.903536131 +0000 UTC m=+318.215733664" observedRunningTime="2026-03-15 00:12:05.471857952 +0000 UTC m=+318.784055495" watchObservedRunningTime="2026-03-15 00:12:05.475854124 +0000 UTC m=+318.788051657" Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.494881 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.500443 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-688bf79b7d-ztr76"] Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.524643 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:12:05 crc kubenswrapper[4861]: I0315 00:12:05.528861 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6c77bfd598-jmrdm"] Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062097 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:06 crc kubenswrapper[4861]: E0315 00:12:06.062596 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a399224b-3b39-4632-bf97-b85c68ca72db" containerName="controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062623 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="a399224b-3b39-4632-bf97-b85c68ca72db" containerName="controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: E0315 00:12:06.062640 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dec95d-ab68-4794-9644-96e48cf6016c" containerName="route-controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062651 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dec95d-ab68-4794-9644-96e48cf6016c" containerName="route-controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: E0315 00:12:06.062666 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062679 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: E0315 00:12:06.062704 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062714 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062890 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062909 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" containerName="oc" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062925 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="a399224b-3b39-4632-bf97-b85c68ca72db" containerName="controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.062937 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dec95d-ab68-4794-9644-96e48cf6016c" containerName="route-controller-manager" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.063664 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.064822 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.065739 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.071304 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.071820 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.072231 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.072360 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.072384 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.072707 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.072951 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.073212 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.073479 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.073973 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.075411 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.076609 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.078654 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.082344 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.091603 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.096835 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.096901 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.128340 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.128427 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.128456 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhngv\" (UniqueName: \"kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.128489 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.128655 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.149510 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.216145 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zz6cr" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="registry-server" probeResult="failure" output=< Mar 15 00:12:06 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:12:06 crc kubenswrapper[4861]: > Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229777 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229837 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229882 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p56r\" (UniqueName: \"kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229937 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229966 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.229995 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.230023 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhngv\" (UniqueName: \"kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.230079 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.230144 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.231845 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.231969 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.232521 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.250453 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.254225 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhngv\" (UniqueName: \"kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv\") pod \"controller-manager-78bbc865f5-p5d22\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.332092 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.332546 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p56r\" (UniqueName: \"kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.332685 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.332868 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.333777 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.334087 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.335994 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.365168 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p56r\" (UniqueName: \"kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r\") pod \"route-controller-manager-869f694dd9-486hd\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.383898 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.392143 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.451332 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.451533 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.510721 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.711802 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:06 crc kubenswrapper[4861]: I0315 00:12:06.859502 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:06 crc kubenswrapper[4861]: W0315 00:12:06.874443 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e33a813_9150_4f43_a114_8e996e5d444b.slice/crio-c62ce625a329728e59cadc4263582be78a2a1b06ceb39d8fbbf51427cb32e6fb WatchSource:0}: Error finding container c62ce625a329728e59cadc4263582be78a2a1b06ceb39d8fbbf51427cb32e6fb: Status 404 returned error can't find the container with id c62ce625a329728e59cadc4263582be78a2a1b06ceb39d8fbbf51427cb32e6fb Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.062054 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.062114 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.417617 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dec95d-ab68-4794-9644-96e48cf6016c" path="/var/lib/kubelet/pods/57dec95d-ab68-4794-9644-96e48cf6016c/volumes" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.419124 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a399224b-3b39-4632-bf97-b85c68ca72db" path="/var/lib/kubelet/pods/a399224b-3b39-4632-bf97-b85c68ca72db/volumes" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.484092 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" event={"ID":"4145cc8b-db85-432f-8ece-08dc7c31317f","Type":"ContainerStarted","Data":"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac"} Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.484178 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" event={"ID":"4145cc8b-db85-432f-8ece-08dc7c31317f","Type":"ContainerStarted","Data":"c6b84a85394c834179210f7d41ac14f6c4d107e2af7596a283a0c5f53d349b46"} Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.485499 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.488046 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" event={"ID":"1e33a813-9150-4f43-a114-8e996e5d444b","Type":"ContainerStarted","Data":"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770"} Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.488093 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" event={"ID":"1e33a813-9150-4f43-a114-8e996e5d444b","Type":"ContainerStarted","Data":"c62ce625a329728e59cadc4263582be78a2a1b06ceb39d8fbbf51427cb32e6fb"} Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.488925 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.497205 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.512814 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" podStartSLOduration=3.512791542 podStartE2EDuration="3.512791542s" podCreationTimestamp="2026-03-15 00:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:12:07.511201638 +0000 UTC m=+320.823399171" watchObservedRunningTime="2026-03-15 00:12:07.512791542 +0000 UTC m=+320.824989065" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.727916 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.752298 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" podStartSLOduration=3.752277101 podStartE2EDuration="3.752277101s" podCreationTimestamp="2026-03-15 00:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:12:07.532304358 +0000 UTC m=+320.844501891" watchObservedRunningTime="2026-03-15 00:12:07.752277101 +0000 UTC m=+321.064474634" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.798763 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:07 crc kubenswrapper[4861]: I0315 00:12:07.798831 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:08 crc kubenswrapper[4861]: I0315 00:12:08.106617 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9447z" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="registry-server" probeResult="failure" output=< Mar 15 00:12:08 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:12:08 crc kubenswrapper[4861]: > Mar 15 00:12:09 crc kubenswrapper[4861]: I0315 00:12:09.054691 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t4p99" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="registry-server" probeResult="failure" output=< Mar 15 00:12:09 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:12:09 crc kubenswrapper[4861]: > Mar 15 00:12:13 crc kubenswrapper[4861]: I0315 00:12:13.923096 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:12:13 crc kubenswrapper[4861]: I0315 00:12:13.969935 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:12:14 crc kubenswrapper[4861]: I0315 00:12:14.404934 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:15 crc kubenswrapper[4861]: I0315 00:12:15.169479 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:12:15 crc kubenswrapper[4861]: I0315 00:12:15.169856 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rkmhv" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="registry-server" containerID="cri-o://01cd5e0d856603c45a8198ddaccfe97a9b6201cf9977c99be55d5a512645baf0" gracePeriod=2 Mar 15 00:12:15 crc kubenswrapper[4861]: I0315 00:12:15.245452 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:12:15 crc kubenswrapper[4861]: I0315 00:12:15.308386 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.152621 4861 generic.go:334] "Generic (PLEG): container finished" podID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerID="01cd5e0d856603c45a8198ddaccfe97a9b6201cf9977c99be55d5a512645baf0" exitCode=0 Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.155152 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerDied","Data":"01cd5e0d856603c45a8198ddaccfe97a9b6201cf9977c99be55d5a512645baf0"} Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.197926 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.400368 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.516832 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.558945 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jtc2\" (UniqueName: \"kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2\") pod \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.559097 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content\") pod \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.559217 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities\") pod \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\" (UID: \"c32c6537-08bb-4067-ad4f-7f2ce276be7b\") " Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.560188 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities" (OuterVolumeSpecName: "utilities") pod "c32c6537-08bb-4067-ad4f-7f2ce276be7b" (UID: "c32c6537-08bb-4067-ad4f-7f2ce276be7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.577805 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2" (OuterVolumeSpecName: "kube-api-access-4jtc2") pod "c32c6537-08bb-4067-ad4f-7f2ce276be7b" (UID: "c32c6537-08bb-4067-ad4f-7f2ce276be7b"). InnerVolumeSpecName "kube-api-access-4jtc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.618853 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c32c6537-08bb-4067-ad4f-7f2ce276be7b" (UID: "c32c6537-08bb-4067-ad4f-7f2ce276be7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.661298 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jtc2\" (UniqueName: \"kubernetes.io/projected/c32c6537-08bb-4067-ad4f-7f2ce276be7b-kube-api-access-4jtc2\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.661338 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:16 crc kubenswrapper[4861]: I0315 00:12:16.661347 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32c6537-08bb-4067-ad4f-7f2ce276be7b-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.132586 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.164671 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rkmhv" event={"ID":"c32c6537-08bb-4067-ad4f-7f2ce276be7b","Type":"ContainerDied","Data":"e267de667edd62e9313009121a42e9ddae018f033650fd7f47d13338c1bf4f73"} Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.164748 4861 scope.go:117] "RemoveContainer" containerID="01cd5e0d856603c45a8198ddaccfe97a9b6201cf9977c99be55d5a512645baf0" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.165082 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rkmhv" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.210712 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.211809 4861 scope.go:117] "RemoveContainer" containerID="835241daf746c1dd07aba17c75ae5fcc2994957383315d9534668e3a456ce6e6" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.215949 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.221145 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rkmhv"] Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.269943 4861 scope.go:117] "RemoveContainer" containerID="bd69696f6a470e446fa7c5784d6076024bd20f5964f89e03951eb305e5a36de1" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.415438 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" path="/var/lib/kubelet/pods/c32c6537-08bb-4067-ad4f-7f2ce276be7b/volumes" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.509934 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j2krj"] Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.846517 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:17 crc kubenswrapper[4861]: I0315 00:12:17.888885 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.365182 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.365515 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6hzjv" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="registry-server" containerID="cri-o://ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc" gracePeriod=2 Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.660699 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.661503 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.664628 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.664807 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.672426 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.706928 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.762897 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.763096 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.765259 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.775257 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.792086 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.812406 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.826274 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.847979 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.856004 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.873253 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.969364 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities\") pod \"9e466e6e-862b-41f2-9639-fa0adb939f84\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.969405 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content\") pod \"9e466e6e-862b-41f2-9639-fa0adb939f84\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.969500 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqkhg\" (UniqueName: \"kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg\") pod \"9e466e6e-862b-41f2-9639-fa0adb939f84\" (UID: \"9e466e6e-862b-41f2-9639-fa0adb939f84\") " Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.978837 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities" (OuterVolumeSpecName: "utilities") pod "9e466e6e-862b-41f2-9639-fa0adb939f84" (UID: "9e466e6e-862b-41f2-9639-fa0adb939f84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:19 crc kubenswrapper[4861]: I0315 00:12:19.978918 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg" (OuterVolumeSpecName: "kube-api-access-cqkhg") pod "9e466e6e-862b-41f2-9639-fa0adb939f84" (UID: "9e466e6e-862b-41f2-9639-fa0adb939f84"). InnerVolumeSpecName "kube-api-access-cqkhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.010083 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e466e6e-862b-41f2-9639-fa0adb939f84" (UID: "9e466e6e-862b-41f2-9639-fa0adb939f84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.073427 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqkhg\" (UniqueName: \"kubernetes.io/projected/9e466e6e-862b-41f2-9639-fa0adb939f84-kube-api-access-cqkhg\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.073465 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.073476 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e466e6e-862b-41f2-9639-fa0adb939f84-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.189310 4861 generic.go:334] "Generic (PLEG): container finished" podID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerID="ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc" exitCode=0 Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.189378 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerDied","Data":"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc"} Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.189428 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6hzjv" event={"ID":"9e466e6e-862b-41f2-9639-fa0adb939f84","Type":"ContainerDied","Data":"e0385562efc1af970c3c699419627f4e0269c93cec174c2b4cd2fd6e0ed5c3fb"} Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.189449 4861 scope.go:117] "RemoveContainer" containerID="ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.189444 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6hzjv" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.222893 4861 scope.go:117] "RemoveContainer" containerID="74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.230542 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.238249 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6hzjv"] Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.247400 4861 scope.go:117] "RemoveContainer" containerID="83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.277830 4861 scope.go:117] "RemoveContainer" containerID="ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc" Mar 15 00:12:20 crc kubenswrapper[4861]: E0315 00:12:20.280135 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc\": container with ID starting with ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc not found: ID does not exist" containerID="ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.280197 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc"} err="failed to get container status \"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc\": rpc error: code = NotFound desc = could not find container \"ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc\": container with ID starting with ba18b0a997d61f10bff2f36a4c60a3f6c867f92a7733b50d91931828cdedd4fc not found: ID does not exist" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.280231 4861 scope.go:117] "RemoveContainer" containerID="74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719" Mar 15 00:12:20 crc kubenswrapper[4861]: E0315 00:12:20.281995 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719\": container with ID starting with 74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719 not found: ID does not exist" containerID="74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.282034 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719"} err="failed to get container status \"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719\": rpc error: code = NotFound desc = could not find container \"74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719\": container with ID starting with 74a0698cb6ea07f6e0a98c5072e801604f4c85664a480733ac2b1a54a8521719 not found: ID does not exist" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.282065 4861 scope.go:117] "RemoveContainer" containerID="83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2" Mar 15 00:12:20 crc kubenswrapper[4861]: E0315 00:12:20.282258 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2\": container with ID starting with 83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2 not found: ID does not exist" containerID="83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2" Mar 15 00:12:20 crc kubenswrapper[4861]: I0315 00:12:20.282282 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2"} err="failed to get container status \"83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2\": rpc error: code = NotFound desc = could not find container \"83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2\": container with ID starting with 83c525af3e811d73e1301fc77a3f59ae6685fc2aa28bab61c088057cf31f2ce2 not found: ID does not exist" Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.198806 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"08ab26b10488cc535b1d75fd30c1be6fd147973e72129c1b7921e26763c01d68"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.199723 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8d47cf7b128dcd976462046fdcbfe93e2dd33f5b6b40aa74e6be1140fe60fdbf"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.207449 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c66d8be2f1786a55a56228bf92b8229fd9ea93e149d140a733b8c71c7d37947d"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.207512 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2006766fe11916f74d2f538e05677e9fdb32b0bb4371d72ef64b6608dd5489c4"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.212941 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3a983a1927d00acbee33050ae4222c3d4057d67c1475f6e4af466ad8612ec8d7"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.213034 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9c91e08865b07dad402c7529348eceeee0d1bd489d00ead6d0260f2852fbfb57"} Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.213247 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.417011 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" path="/var/lib/kubelet/pods/9e466e6e-862b-41f2-9639-fa0adb939f84/volumes" Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.565720 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:12:21 crc kubenswrapper[4861]: I0315 00:12:21.566035 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t4p99" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="registry-server" containerID="cri-o://7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847" gracePeriod=2 Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.107008 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.205236 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities\") pod \"c842b914-5c03-4d7e-85f5-fae121eec542\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.206058 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content\") pod \"c842b914-5c03-4d7e-85f5-fae121eec542\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.206502 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh8k4\" (UniqueName: \"kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4\") pod \"c842b914-5c03-4d7e-85f5-fae121eec542\" (UID: \"c842b914-5c03-4d7e-85f5-fae121eec542\") " Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.209062 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities" (OuterVolumeSpecName: "utilities") pod "c842b914-5c03-4d7e-85f5-fae121eec542" (UID: "c842b914-5c03-4d7e-85f5-fae121eec542"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.215914 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4" (OuterVolumeSpecName: "kube-api-access-mh8k4") pod "c842b914-5c03-4d7e-85f5-fae121eec542" (UID: "c842b914-5c03-4d7e-85f5-fae121eec542"). InnerVolumeSpecName "kube-api-access-mh8k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.220923 4861 generic.go:334] "Generic (PLEG): container finished" podID="c842b914-5c03-4d7e-85f5-fae121eec542" containerID="7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847" exitCode=0 Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.221368 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4p99" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.221374 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerDied","Data":"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847"} Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.221646 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4p99" event={"ID":"c842b914-5c03-4d7e-85f5-fae121eec542","Type":"ContainerDied","Data":"8fcf7d546d42f7655b97c2d4c7cb8032f11c5a5dad17280b04889c2fda689c7f"} Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.221744 4861 scope.go:117] "RemoveContainer" containerID="7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.265266 4861 scope.go:117] "RemoveContainer" containerID="d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.289857 4861 scope.go:117] "RemoveContainer" containerID="fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.309119 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh8k4\" (UniqueName: \"kubernetes.io/projected/c842b914-5c03-4d7e-85f5-fae121eec542-kube-api-access-mh8k4\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.309152 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.311319 4861 scope.go:117] "RemoveContainer" containerID="7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847" Mar 15 00:12:22 crc kubenswrapper[4861]: E0315 00:12:22.311811 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847\": container with ID starting with 7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847 not found: ID does not exist" containerID="7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.311897 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847"} err="failed to get container status \"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847\": rpc error: code = NotFound desc = could not find container \"7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847\": container with ID starting with 7fddaa97c9b13e754237cc8f842883d7629d4589f55ffb5482d9f60a82b45847 not found: ID does not exist" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.311958 4861 scope.go:117] "RemoveContainer" containerID="d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43" Mar 15 00:12:22 crc kubenswrapper[4861]: E0315 00:12:22.312401 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43\": container with ID starting with d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43 not found: ID does not exist" containerID="d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.312446 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43"} err="failed to get container status \"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43\": rpc error: code = NotFound desc = could not find container \"d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43\": container with ID starting with d542c3983b0fada3a8c0a58d50e8dc14569f3c78c4a7dbbd134e5c41db198a43 not found: ID does not exist" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.312466 4861 scope.go:117] "RemoveContainer" containerID="fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65" Mar 15 00:12:22 crc kubenswrapper[4861]: E0315 00:12:22.312880 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65\": container with ID starting with fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65 not found: ID does not exist" containerID="fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.312914 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65"} err="failed to get container status \"fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65\": rpc error: code = NotFound desc = could not find container \"fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65\": container with ID starting with fa84cc5c404ff853d92fb94b7e263e2518347c4e1c9dbc92a103369f99596d65 not found: ID does not exist" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.341097 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c842b914-5c03-4d7e-85f5-fae121eec542" (UID: "c842b914-5c03-4d7e-85f5-fae121eec542"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.410697 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c842b914-5c03-4d7e-85f5-fae121eec542-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.572811 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:12:22 crc kubenswrapper[4861]: I0315 00:12:22.579773 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t4p99"] Mar 15 00:12:23 crc kubenswrapper[4861]: I0315 00:12:23.418596 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" path="/var/lib/kubelet/pods/c842b914-5c03-4d7e-85f5-fae121eec542/volumes" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.155939 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.156173 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" podUID="1e33a813-9150-4f43-a114-8e996e5d444b" containerName="controller-manager" containerID="cri-o://970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770" gracePeriod=30 Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.281100 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.281381 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" podUID="4145cc8b-db85-432f-8ece-08dc7c31317f" containerName="route-controller-manager" containerID="cri-o://56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac" gracePeriod=30 Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.899650 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.905460 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.946984 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert\") pod \"4145cc8b-db85-432f-8ece-08dc7c31317f\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947108 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p56r\" (UniqueName: \"kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r\") pod \"4145cc8b-db85-432f-8ece-08dc7c31317f\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947203 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config\") pod \"1e33a813-9150-4f43-a114-8e996e5d444b\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947230 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca\") pod \"4145cc8b-db85-432f-8ece-08dc7c31317f\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947365 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca\") pod \"1e33a813-9150-4f43-a114-8e996e5d444b\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947408 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config\") pod \"4145cc8b-db85-432f-8ece-08dc7c31317f\" (UID: \"4145cc8b-db85-432f-8ece-08dc7c31317f\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947438 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhngv\" (UniqueName: \"kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv\") pod \"1e33a813-9150-4f43-a114-8e996e5d444b\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947529 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles\") pod \"1e33a813-9150-4f43-a114-8e996e5d444b\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.947582 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert\") pod \"1e33a813-9150-4f43-a114-8e996e5d444b\" (UID: \"1e33a813-9150-4f43-a114-8e996e5d444b\") " Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.948840 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca" (OuterVolumeSpecName: "client-ca") pod "4145cc8b-db85-432f-8ece-08dc7c31317f" (UID: "4145cc8b-db85-432f-8ece-08dc7c31317f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.948931 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config" (OuterVolumeSpecName: "config") pod "4145cc8b-db85-432f-8ece-08dc7c31317f" (UID: "4145cc8b-db85-432f-8ece-08dc7c31317f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.949371 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1e33a813-9150-4f43-a114-8e996e5d444b" (UID: "1e33a813-9150-4f43-a114-8e996e5d444b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.949535 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca" (OuterVolumeSpecName: "client-ca") pod "1e33a813-9150-4f43-a114-8e996e5d444b" (UID: "1e33a813-9150-4f43-a114-8e996e5d444b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.951341 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config" (OuterVolumeSpecName: "config") pod "1e33a813-9150-4f43-a114-8e996e5d444b" (UID: "1e33a813-9150-4f43-a114-8e996e5d444b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.956888 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1e33a813-9150-4f43-a114-8e996e5d444b" (UID: "1e33a813-9150-4f43-a114-8e996e5d444b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.957110 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r" (OuterVolumeSpecName: "kube-api-access-4p56r") pod "4145cc8b-db85-432f-8ece-08dc7c31317f" (UID: "4145cc8b-db85-432f-8ece-08dc7c31317f"). InnerVolumeSpecName "kube-api-access-4p56r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.959153 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv" (OuterVolumeSpecName: "kube-api-access-mhngv") pod "1e33a813-9150-4f43-a114-8e996e5d444b" (UID: "1e33a813-9150-4f43-a114-8e996e5d444b"). InnerVolumeSpecName "kube-api-access-mhngv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:24 crc kubenswrapper[4861]: I0315 00:12:24.959535 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4145cc8b-db85-432f-8ece-08dc7c31317f" (UID: "4145cc8b-db85-432f-8ece-08dc7c31317f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048489 4861 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048537 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e33a813-9150-4f43-a114-8e996e5d444b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048553 4861 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4145cc8b-db85-432f-8ece-08dc7c31317f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048586 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p56r\" (UniqueName: \"kubernetes.io/projected/4145cc8b-db85-432f-8ece-08dc7c31317f-kube-api-access-4p56r\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048603 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048615 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048628 4861 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e33a813-9150-4f43-a114-8e996e5d444b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048640 4861 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4145cc8b-db85-432f-8ece-08dc7c31317f-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.048653 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhngv\" (UniqueName: \"kubernetes.io/projected/1e33a813-9150-4f43-a114-8e996e5d444b-kube-api-access-mhngv\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.246494 4861 generic.go:334] "Generic (PLEG): container finished" podID="4145cc8b-db85-432f-8ece-08dc7c31317f" containerID="56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac" exitCode=0 Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.246602 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" event={"ID":"4145cc8b-db85-432f-8ece-08dc7c31317f","Type":"ContainerDied","Data":"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac"} Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.246677 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" event={"ID":"4145cc8b-db85-432f-8ece-08dc7c31317f","Type":"ContainerDied","Data":"c6b84a85394c834179210f7d41ac14f6c4d107e2af7596a283a0c5f53d349b46"} Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.246715 4861 scope.go:117] "RemoveContainer" containerID="56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.247401 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.248783 4861 generic.go:334] "Generic (PLEG): container finished" podID="1e33a813-9150-4f43-a114-8e996e5d444b" containerID="970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770" exitCode=0 Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.248828 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" event={"ID":"1e33a813-9150-4f43-a114-8e996e5d444b","Type":"ContainerDied","Data":"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770"} Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.248855 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" event={"ID":"1e33a813-9150-4f43-a114-8e996e5d444b","Type":"ContainerDied","Data":"c62ce625a329728e59cadc4263582be78a2a1b06ceb39d8fbbf51427cb32e6fb"} Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.249020 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-78bbc865f5-p5d22" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.276941 4861 scope.go:117] "RemoveContainer" containerID="56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac" Mar 15 00:12:25 crc kubenswrapper[4861]: E0315 00:12:25.277759 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac\": container with ID starting with 56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac not found: ID does not exist" containerID="56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.277800 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac"} err="failed to get container status \"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac\": rpc error: code = NotFound desc = could not find container \"56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac\": container with ID starting with 56358238cd6365d6ed24216fc06d9c4fd392dac7b55e1f5be152586bd1f152ac not found: ID does not exist" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.277825 4861 scope.go:117] "RemoveContainer" containerID="970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.338724 4861 scope.go:117] "RemoveContainer" containerID="970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770" Mar 15 00:12:25 crc kubenswrapper[4861]: E0315 00:12:25.340547 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770\": container with ID starting with 970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770 not found: ID does not exist" containerID="970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.341231 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770"} err="failed to get container status \"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770\": rpc error: code = NotFound desc = could not find container \"970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770\": container with ID starting with 970123b14ec7c51636c989a2692044dded130a9a5a12d0c2988ff8434854b770 not found: ID does not exist" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.343667 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.348170 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-78bbc865f5-p5d22"] Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.363542 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.368006 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-869f694dd9-486hd"] Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.427954 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e33a813-9150-4f43-a114-8e996e5d444b" path="/var/lib/kubelet/pods/1e33a813-9150-4f43-a114-8e996e5d444b/volumes" Mar 15 00:12:25 crc kubenswrapper[4861]: I0315 00:12:25.431888 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4145cc8b-db85-432f-8ece-08dc7c31317f" path="/var/lib/kubelet/pods/4145cc8b-db85-432f-8ece-08dc7c31317f/volumes" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.084128 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49"] Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085014 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085029 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085041 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e33a813-9150-4f43-a114-8e996e5d444b" containerName="controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085049 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e33a813-9150-4f43-a114-8e996e5d444b" containerName="controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085058 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4145cc8b-db85-432f-8ece-08dc7c31317f" containerName="route-controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085066 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="4145cc8b-db85-432f-8ece-08dc7c31317f" containerName="route-controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085074 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085080 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085091 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085097 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085109 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085115 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085126 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085132 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085142 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085148 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="extract-utilities" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085159 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085164 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="extract-content" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085172 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085178 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: E0315 00:12:26.085184 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085190 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085293 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32c6537-08bb-4067-ad4f-7f2ce276be7b" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085306 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e466e6e-862b-41f2-9639-fa0adb939f84" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085315 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="c842b914-5c03-4d7e-85f5-fae121eec542" containerName="registry-server" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085325 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="4145cc8b-db85-432f-8ece-08dc7c31317f" containerName="route-controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085335 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e33a813-9150-4f43-a114-8e996e5d444b" containerName="controller-manager" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.085843 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.086457 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-854c47445c-5zjjv"] Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.087395 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.090184 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.090507 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.090943 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091454 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091581 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091733 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091765 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091862 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.091951 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.092104 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.092129 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.098046 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.101913 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49"] Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.104998 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.110173 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-854c47445c-5zjjv"] Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164000 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-proxy-ca-bundles\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164087 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-config\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164177 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-client-ca\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164207 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c53026aa-83d4-41b7-95aa-03cf53e42543-serving-cert\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164238 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a93fa3ba-4305-4e31-bb11-0060f53bd762-serving-cert\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164272 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwzqz\" (UniqueName: \"kubernetes.io/projected/c53026aa-83d4-41b7-95aa-03cf53e42543-kube-api-access-xwzqz\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164315 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj77n\" (UniqueName: \"kubernetes.io/projected/a93fa3ba-4305-4e31-bb11-0060f53bd762-kube-api-access-jj77n\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164371 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-config\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.164400 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-client-ca\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265133 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a93fa3ba-4305-4e31-bb11-0060f53bd762-serving-cert\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265197 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c53026aa-83d4-41b7-95aa-03cf53e42543-serving-cert\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265222 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwzqz\" (UniqueName: \"kubernetes.io/projected/c53026aa-83d4-41b7-95aa-03cf53e42543-kube-api-access-xwzqz\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265249 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj77n\" (UniqueName: \"kubernetes.io/projected/a93fa3ba-4305-4e31-bb11-0060f53bd762-kube-api-access-jj77n\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265276 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-config\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265293 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-client-ca\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265314 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-proxy-ca-bundles\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265347 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-config\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.265378 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-client-ca\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.266895 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-client-ca\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.267683 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-client-ca\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.267785 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c53026aa-83d4-41b7-95aa-03cf53e42543-config\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.268263 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-config\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.269615 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a93fa3ba-4305-4e31-bb11-0060f53bd762-proxy-ca-bundles\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.272233 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a93fa3ba-4305-4e31-bb11-0060f53bd762-serving-cert\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.285264 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c53026aa-83d4-41b7-95aa-03cf53e42543-serving-cert\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.286003 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj77n\" (UniqueName: \"kubernetes.io/projected/a93fa3ba-4305-4e31-bb11-0060f53bd762-kube-api-access-jj77n\") pod \"controller-manager-854c47445c-5zjjv\" (UID: \"a93fa3ba-4305-4e31-bb11-0060f53bd762\") " pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.291063 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwzqz\" (UniqueName: \"kubernetes.io/projected/c53026aa-83d4-41b7-95aa-03cf53e42543-kube-api-access-xwzqz\") pod \"route-controller-manager-55c6c4875d-vnr49\" (UID: \"c53026aa-83d4-41b7-95aa-03cf53e42543\") " pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.457682 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.469047 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:26 crc kubenswrapper[4861]: I0315 00:12:26.984643 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-854c47445c-5zjjv"] Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.037727 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49"] Mar 15 00:12:27 crc kubenswrapper[4861]: W0315 00:12:27.070260 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc53026aa_83d4_41b7_95aa_03cf53e42543.slice/crio-8ab234c02ccf629fcceffa329f59d2f1af208cded5c3d2ab8536c65cf0a64be5 WatchSource:0}: Error finding container 8ab234c02ccf629fcceffa329f59d2f1af208cded5c3d2ab8536c65cf0a64be5: Status 404 returned error can't find the container with id 8ab234c02ccf629fcceffa329f59d2f1af208cded5c3d2ab8536c65cf0a64be5 Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.264635 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" event={"ID":"a93fa3ba-4305-4e31-bb11-0060f53bd762","Type":"ContainerStarted","Data":"814c26b4d819b040b48358586ca108f3ae8d897c0f0e2a1b570d06720ca6c4cd"} Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.265181 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" event={"ID":"a93fa3ba-4305-4e31-bb11-0060f53bd762","Type":"ContainerStarted","Data":"654ab47c078552c2384a30317d7b168e727b85d02cce8e7f48cf9afdbb81b0a5"} Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.266806 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.268971 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" event={"ID":"c53026aa-83d4-41b7-95aa-03cf53e42543","Type":"ContainerStarted","Data":"ffe810ac60d60b18e1085ae0fc6ff1303b098847185a3ab814a51afe3819c336"} Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.269103 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" event={"ID":"c53026aa-83d4-41b7-95aa-03cf53e42543","Type":"ContainerStarted","Data":"8ab234c02ccf629fcceffa329f59d2f1af208cded5c3d2ab8536c65cf0a64be5"} Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.269184 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.280738 4861 patch_prober.go:28] interesting pod/route-controller-manager-55c6c4875d-vnr49 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.280796 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" podUID="c53026aa-83d4-41b7-95aa-03cf53e42543" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.285147 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.308700 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-854c47445c-5zjjv" podStartSLOduration=3.308663289 podStartE2EDuration="3.308663289s" podCreationTimestamp="2026-03-15 00:12:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:12:27.305197882 +0000 UTC m=+340.617395455" watchObservedRunningTime="2026-03-15 00:12:27.308663289 +0000 UTC m=+340.620860862" Mar 15 00:12:27 crc kubenswrapper[4861]: I0315 00:12:27.333397 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" podStartSLOduration=3.333356112 podStartE2EDuration="3.333356112s" podCreationTimestamp="2026-03-15 00:12:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:12:27.331607053 +0000 UTC m=+340.643804666" watchObservedRunningTime="2026-03-15 00:12:27.333356112 +0000 UTC m=+340.645553685" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.282358 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55c6c4875d-vnr49" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.350877 4861 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.351663 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.351756 4861 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.352350 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307" gracePeriod=15 Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.352412 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9" gracePeriod=15 Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.352439 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281" gracePeriod=15 Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.352467 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091" gracePeriod=15 Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.352354 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667" gracePeriod=15 Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.353834 4861 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354151 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354172 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354196 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354210 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354230 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354244 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354360 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354403 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354417 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354430 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354447 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354460 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354477 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354488 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354505 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.354517 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.354542 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355190 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355405 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355435 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355452 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355471 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355492 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355519 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355551 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 15 00:12:28 crc kubenswrapper[4861]: E0315 00:12:28.355820 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.355841 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.356076 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.356099 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.399106 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.399665 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.399715 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.399743 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.399911 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501308 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501383 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501423 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501451 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501471 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501489 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501512 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501605 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501609 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501496 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501658 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501623 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.501496 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.602910 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.603016 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.603061 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.603019 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.603128 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:28 crc kubenswrapper[4861]: I0315 00:12:28.603239 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.287069 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.289662 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.290797 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307" exitCode=0 Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.290856 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091" exitCode=0 Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.290872 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9" exitCode=0 Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.290893 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281" exitCode=2 Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.290988 4861 scope.go:117] "RemoveContainer" containerID="dd960779840eac04139cbd5d259780c453001bd11b610d954606318ddf929940" Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.294020 4861 generic.go:334] "Generic (PLEG): container finished" podID="7d821af9-853a-43b8-8a31-c7df6895c19a" containerID="13e60e578154bdfac6c323436f2ebd375b86188e7a57c4fb1cdba0c4c1c755bd" exitCode=0 Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.294072 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7d821af9-853a-43b8-8a31-c7df6895c19a","Type":"ContainerDied","Data":"13e60e578154bdfac6c323436f2ebd375b86188e7a57c4fb1cdba0c4c1c755bd"} Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.296203 4861 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:29 crc kubenswrapper[4861]: I0315 00:12:29.296853 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.317090 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.772968 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.774495 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.774716 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.775200 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.775775 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.776089 4861 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886179 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886287 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock\") pod \"7d821af9-853a-43b8-8a31-c7df6895c19a\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886339 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886380 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886446 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access\") pod \"7d821af9-853a-43b8-8a31-c7df6895c19a\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886480 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir\") pod \"7d821af9-853a-43b8-8a31-c7df6895c19a\" (UID: \"7d821af9-853a-43b8-8a31-c7df6895c19a\") " Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886963 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886988 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.886979 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock" (OuterVolumeSpecName: "var-lock") pod "7d821af9-853a-43b8-8a31-c7df6895c19a" (UID: "7d821af9-853a-43b8-8a31-c7df6895c19a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.887029 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.887146 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7d821af9-853a-43b8-8a31-c7df6895c19a" (UID: "7d821af9-853a-43b8-8a31-c7df6895c19a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.887900 4861 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.888034 4861 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.888132 4861 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.888214 4861 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7d821af9-853a-43b8-8a31-c7df6895c19a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.888294 4861 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.895229 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7d821af9-853a-43b8-8a31-c7df6895c19a" (UID: "7d821af9-853a-43b8-8a31-c7df6895c19a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:30 crc kubenswrapper[4861]: I0315 00:12:30.990308 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d821af9-853a-43b8-8a31-c7df6895c19a-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.340349 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.341823 4861 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667" exitCode=0 Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.341931 4861 scope.go:117] "RemoveContainer" containerID="4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.342062 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.347197 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7d821af9-853a-43b8-8a31-c7df6895c19a","Type":"ContainerDied","Data":"173f6b46cba544afa2e76cbbe0dd44c46a7f616b5e7ed8fff497e7ea3a8251d2"} Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.347281 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="173f6b46cba544afa2e76cbbe0dd44c46a7f616b5e7ed8fff497e7ea3a8251d2" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.347404 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.385872 4861 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.386470 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.386978 4861 scope.go:117] "RemoveContainer" containerID="73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.387614 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.388169 4861 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.424370 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.427178 4861 scope.go:117] "RemoveContainer" containerID="d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.453582 4861 scope.go:117] "RemoveContainer" containerID="91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.473446 4861 scope.go:117] "RemoveContainer" containerID="c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.502471 4861 scope.go:117] "RemoveContainer" containerID="cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.531915 4861 scope.go:117] "RemoveContainer" containerID="4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.533196 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\": container with ID starting with 4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307 not found: ID does not exist" containerID="4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.533252 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307"} err="failed to get container status \"4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\": rpc error: code = NotFound desc = could not find container \"4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307\": container with ID starting with 4dd54e94f49cd30fe90cfa1f71233bd0b7e1658e6953c4d64e75e1ba7b72e307 not found: ID does not exist" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.533288 4861 scope.go:117] "RemoveContainer" containerID="73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.533766 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\": container with ID starting with 73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091 not found: ID does not exist" containerID="73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.533832 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091"} err="failed to get container status \"73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\": rpc error: code = NotFound desc = could not find container \"73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091\": container with ID starting with 73741f29e87ea0838d7af28d198f2fe883999fcd46553ea9c6607067c2bb2091 not found: ID does not exist" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.533877 4861 scope.go:117] "RemoveContainer" containerID="d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.534373 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\": container with ID starting with d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9 not found: ID does not exist" containerID="d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.534423 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9"} err="failed to get container status \"d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\": rpc error: code = NotFound desc = could not find container \"d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9\": container with ID starting with d2424aa858668e94fd300104a6c7fb574cdc3a252f286919354ba1e4061004c9 not found: ID does not exist" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.534456 4861 scope.go:117] "RemoveContainer" containerID="91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.534839 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\": container with ID starting with 91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281 not found: ID does not exist" containerID="91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.534878 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281"} err="failed to get container status \"91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\": rpc error: code = NotFound desc = could not find container \"91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281\": container with ID starting with 91d0e526eeff29a08548de51c50bf46b715547e23a35882faba6c772e58a9281 not found: ID does not exist" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.534900 4861 scope.go:117] "RemoveContainer" containerID="c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.536213 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\": container with ID starting with c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667 not found: ID does not exist" containerID="c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.536260 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667"} err="failed to get container status \"c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\": rpc error: code = NotFound desc = could not find container \"c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667\": container with ID starting with c97db08e12d17b1b0c6e6fbb9b2c4d939e66efceb8f094a4395e943e106be667 not found: ID does not exist" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.536288 4861 scope.go:117] "RemoveContainer" containerID="cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964" Mar 15 00:12:31 crc kubenswrapper[4861]: E0315 00:12:31.536998 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\": container with ID starting with cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964 not found: ID does not exist" containerID="cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964" Mar 15 00:12:31 crc kubenswrapper[4861]: I0315 00:12:31.537028 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964"} err="failed to get container status \"cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\": rpc error: code = NotFound desc = could not find container \"cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964\": container with ID starting with cbc1ffa64858b4feaafa1a30a23c33fc9d302d684c58d668c905f8abce8c8964 not found: ID does not exist" Mar 15 00:12:33 crc kubenswrapper[4861]: E0315 00:12:33.401882 4861 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:33 crc kubenswrapper[4861]: I0315 00:12:33.402834 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:33 crc kubenswrapper[4861]: W0315 00:12:33.427786 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-5f3d04c42e5681b203216b30e47acc06324ba5d9c529e4768ada713f456eeead WatchSource:0}: Error finding container 5f3d04c42e5681b203216b30e47acc06324ba5d9c529e4768ada713f456eeead: Status 404 returned error can't find the container with id 5f3d04c42e5681b203216b30e47acc06324ba5d9c529e4768ada713f456eeead Mar 15 00:12:33 crc kubenswrapper[4861]: E0315 00:12:33.431199 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189cdb5ec972b55a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:12:33.430279514 +0000 UTC m=+346.742477057,LastTimestamp:2026-03-15 00:12:33.430279514 +0000 UTC m=+346.742477057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:12:34 crc kubenswrapper[4861]: I0315 00:12:34.379534 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"eb3b181c483ec061d4b716a4f46ff5bc65b7c1f79a65bf5ef0a8beb64eeb0f76"} Mar 15 00:12:34 crc kubenswrapper[4861]: I0315 00:12:34.380550 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"5f3d04c42e5681b203216b30e47acc06324ba5d9c529e4768ada713f456eeead"} Mar 15 00:12:34 crc kubenswrapper[4861]: E0315 00:12:34.381677 4861 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:12:34 crc kubenswrapper[4861]: I0315 00:12:34.382213 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:34 crc kubenswrapper[4861]: E0315 00:12:34.443461 4861 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" volumeName="registry-storage" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.603697 4861 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.604478 4861 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.605436 4861 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.605994 4861 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.606502 4861 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:36 crc kubenswrapper[4861]: I0315 00:12:36.606593 4861 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.607364 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 15 00:12:36 crc kubenswrapper[4861]: E0315 00:12:36.808677 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 15 00:12:37 crc kubenswrapper[4861]: E0315 00:12:37.196624 4861 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189cdb5ec972b55a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-15 00:12:33.430279514 +0000 UTC m=+346.742477057,LastTimestamp:2026-03-15 00:12:33.430279514 +0000 UTC m=+346.742477057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 15 00:12:37 crc kubenswrapper[4861]: E0315 00:12:37.210496 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 15 00:12:37 crc kubenswrapper[4861]: I0315 00:12:37.413440 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:38 crc kubenswrapper[4861]: E0315 00:12:38.011238 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 15 00:12:39 crc kubenswrapper[4861]: E0315 00:12:39.612376 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.444949 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.447147 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.447220 4861 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968" exitCode=1 Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.447269 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968"} Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.448075 4861 scope.go:117] "RemoveContainer" containerID="3779a76be1a81a1dc43c471e856423a4cb64277c6e0e53642d3e0037b06d1968" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.448724 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.449210 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:41 crc kubenswrapper[4861]: I0315 00:12:41.912981 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.467135 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.469367 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.469467 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"40f052c1d1ec48e121b442ef31939fd3b024f24eef23b832d5067cd165c55b9f"} Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.471144 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.471773 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:42 crc kubenswrapper[4861]: I0315 00:12:42.546847 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" containerID="cri-o://22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90" gracePeriod=15 Mar 15 00:12:42 crc kubenswrapper[4861]: E0315 00:12:42.813998 4861 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="6.4s" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.166466 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.167170 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.167526 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.168062 4861 status_manager.go:851] "Failed to get status for pod" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-j2krj\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192349 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqhsv\" (UniqueName: \"kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192446 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192504 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192586 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192637 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192706 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192769 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192802 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192823 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192867 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192935 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.192998 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.193031 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.193074 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.193206 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") pod \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\" (UID: \"1e8992d1-2a3f-4a3b-b857-821cfb8bed31\") " Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.193648 4861 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.194052 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.194073 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.195000 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.195506 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.200792 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.201490 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv" (OuterVolumeSpecName: "kube-api-access-nqhsv") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "kube-api-access-nqhsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.201819 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.202168 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.203084 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.203138 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.203526 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.203702 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.205256 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "1e8992d1-2a3f-4a3b-b857-821cfb8bed31" (UID: "1e8992d1-2a3f-4a3b-b857-821cfb8bed31"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295499 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295544 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295576 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqhsv\" (UniqueName: \"kubernetes.io/projected/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-kube-api-access-nqhsv\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295590 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295603 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295621 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295637 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295654 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295668 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295681 4861 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295693 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295706 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.295722 4861 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1e8992d1-2a3f-4a3b-b857-821cfb8bed31-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.408975 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.410310 4861 status_manager.go:851] "Failed to get status for pod" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-j2krj\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.410998 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.411728 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.431994 4861 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.432053 4861 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:43 crc kubenswrapper[4861]: E0315 00:12:43.432747 4861 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.433585 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:43 crc kubenswrapper[4861]: W0315 00:12:43.472094 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b43edbb3dce460ec9f42db4b4d12d788b2d4566ed200f8c70d0d2e9526e24f9f WatchSource:0}: Error finding container b43edbb3dce460ec9f42db4b4d12d788b2d4566ed200f8c70d0d2e9526e24f9f: Status 404 returned error can't find the container with id b43edbb3dce460ec9f42db4b4d12d788b2d4566ed200f8c70d0d2e9526e24f9f Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.480965 4861 generic.go:334] "Generic (PLEG): container finished" podID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerID="22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90" exitCode=0 Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.481059 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.481062 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" event={"ID":"1e8992d1-2a3f-4a3b-b857-821cfb8bed31","Type":"ContainerDied","Data":"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90"} Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.481160 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" event={"ID":"1e8992d1-2a3f-4a3b-b857-821cfb8bed31","Type":"ContainerDied","Data":"cb6ebb91900b2fad5eec9fcfeb9e366b28505cf7954d8169e2313ffebd95d69b"} Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.481195 4861 scope.go:117] "RemoveContainer" containerID="22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.481919 4861 status_manager.go:851] "Failed to get status for pod" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-j2krj\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.482401 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.483285 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.489348 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.489829 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.490280 4861 status_manager.go:851] "Failed to get status for pod" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-j2krj\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.521096 4861 scope.go:117] "RemoveContainer" containerID="22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90" Mar 15 00:12:43 crc kubenswrapper[4861]: E0315 00:12:43.522056 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90\": container with ID starting with 22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90 not found: ID does not exist" containerID="22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90" Mar 15 00:12:43 crc kubenswrapper[4861]: I0315 00:12:43.522144 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90"} err="failed to get container status \"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90\": rpc error: code = NotFound desc = could not find container \"22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90\": container with ID starting with 22515914d75ddeef37bc78a2c2c9159d2e1f53f446cd2fefd4c9178d746b6a90 not found: ID does not exist" Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.498352 4861 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="2edc1c708c920d753f11f4bcc92c7886ebc3dc180dbf6810a2a59b3de0c0aeaf" exitCode=0 Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.498590 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"2edc1c708c920d753f11f4bcc92c7886ebc3dc180dbf6810a2a59b3de0c0aeaf"} Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.498962 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b43edbb3dce460ec9f42db4b4d12d788b2d4566ed200f8c70d0d2e9526e24f9f"} Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.499539 4861 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.499953 4861 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.500238 4861 status_manager.go:851] "Failed to get status for pod" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" pod="openshift-authentication/oauth-openshift-558db77b4-j2krj" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-j2krj\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.500666 4861 status_manager.go:851] "Failed to get status for pod" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:44 crc kubenswrapper[4861]: E0315 00:12:44.500870 4861 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:44 crc kubenswrapper[4861]: I0315 00:12:44.501003 4861 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 15 00:12:45 crc kubenswrapper[4861]: I0315 00:12:45.513975 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"03f08148c42bb69c6f0a25418cd859231439793d706112076c25e887e3f5bda9"} Mar 15 00:12:45 crc kubenswrapper[4861]: I0315 00:12:45.514805 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dd62826e12171e9e96559701400ae63180f3de7135c4d9626916fcf73b28d8ba"} Mar 15 00:12:45 crc kubenswrapper[4861]: I0315 00:12:45.514903 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dffe241a5b4013b95e2dda5dcb764e018d0b7ba1c0bcd9fd08ebdf43007ca72c"} Mar 15 00:12:46 crc kubenswrapper[4861]: I0315 00:12:46.523749 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1dd95b81b6fcedb8fff65d14a6cd6132047d40b9349ceb5500b76e4c403added"} Mar 15 00:12:46 crc kubenswrapper[4861]: I0315 00:12:46.524210 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"af7214e3610ca354925cf32752557728cbf116b0b40c382c2cce47108a6eaee3"} Mar 15 00:12:46 crc kubenswrapper[4861]: I0315 00:12:46.524298 4861 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:46 crc kubenswrapper[4861]: I0315 00:12:46.524330 4861 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:46 crc kubenswrapper[4861]: I0315 00:12:46.524487 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:47 crc kubenswrapper[4861]: I0315 00:12:47.005012 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:12:47 crc kubenswrapper[4861]: I0315 00:12:47.010747 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:12:47 crc kubenswrapper[4861]: I0315 00:12:47.529331 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:12:48 crc kubenswrapper[4861]: I0315 00:12:48.434530 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:48 crc kubenswrapper[4861]: I0315 00:12:48.434662 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:48 crc kubenswrapper[4861]: I0315 00:12:48.446621 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:51 crc kubenswrapper[4861]: I0315 00:12:51.539756 4861 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:12:51 crc kubenswrapper[4861]: I0315 00:12:51.706447 4861 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0792ebff-522b-4ef2-8519-6a265c7257c7" Mar 15 00:12:52 crc kubenswrapper[4861]: I0315 00:12:52.570009 4861 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:52 crc kubenswrapper[4861]: I0315 00:12:52.570594 4861 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:12:52 crc kubenswrapper[4861]: I0315 00:12:52.574318 4861 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0792ebff-522b-4ef2-8519-6a265c7257c7" Mar 15 00:12:55 crc kubenswrapper[4861]: I0315 00:12:55.279331 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:12:55 crc kubenswrapper[4861]: I0315 00:12:55.282963 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 15 00:12:55 crc kubenswrapper[4861]: I0315 00:12:55.298439 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/822fe19d-7cd0-44c8-b1ef-374b8a245328-metrics-certs\") pod \"network-metrics-daemon-t4ghv\" (UID: \"822fe19d-7cd0-44c8-b1ef-374b8a245328\") " pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:12:55 crc kubenswrapper[4861]: I0315 00:12:55.539757 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 15 00:12:55 crc kubenswrapper[4861]: I0315 00:12:55.547820 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-t4ghv" Mar 15 00:12:56 crc kubenswrapper[4861]: I0315 00:12:56.606588 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" event={"ID":"822fe19d-7cd0-44c8-b1ef-374b8a245328","Type":"ContainerStarted","Data":"499b9ccea6eb51f1d60aae50b40fd8e426d4f8436380d152b5f543e7ab0e42d7"} Mar 15 00:12:56 crc kubenswrapper[4861]: I0315 00:12:56.606981 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" event={"ID":"822fe19d-7cd0-44c8-b1ef-374b8a245328","Type":"ContainerStarted","Data":"3eca7fb711ecaba510f6425b54287cd92e91c4d7f6f5f6f5a8bb35a502351be0"} Mar 15 00:12:57 crc kubenswrapper[4861]: I0315 00:12:57.619954 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-t4ghv" event={"ID":"822fe19d-7cd0-44c8-b1ef-374b8a245328","Type":"ContainerStarted","Data":"f9723baf90e8fab206f1f2d13bb4928b569f35f787c1b866b150102d145afe3b"} Mar 15 00:12:59 crc kubenswrapper[4861]: I0315 00:12:59.648111 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 15 00:12:59 crc kubenswrapper[4861]: I0315 00:12:59.862486 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 15 00:13:01 crc kubenswrapper[4861]: I0315 00:13:01.185337 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 15 00:13:01 crc kubenswrapper[4861]: I0315 00:13:01.374482 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 15 00:13:01 crc kubenswrapper[4861]: I0315 00:13:01.420201 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 15 00:13:02 crc kubenswrapper[4861]: I0315 00:13:02.415983 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 15 00:13:02 crc kubenswrapper[4861]: I0315 00:13:02.465633 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 15 00:13:02 crc kubenswrapper[4861]: I0315 00:13:02.708707 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.165962 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.242781 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.366816 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.584026 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.725132 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.744192 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.777439 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 15 00:13:03 crc kubenswrapper[4861]: I0315 00:13:03.865070 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.148506 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.166659 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.194299 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.354875 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.360372 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.403948 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.430967 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.577609 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.749653 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.757911 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.785894 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.814370 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.846416 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 15 00:13:04 crc kubenswrapper[4861]: I0315 00:13:04.973001 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.009325 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.014498 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.022965 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.060408 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.085673 4861 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.115922 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.151428 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.215991 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.253363 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.388024 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.424067 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.473019 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.679418 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.715414 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.816974 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.836381 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 15 00:13:05 crc kubenswrapper[4861]: I0315 00:13:05.850026 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.025418 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.056394 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.079248 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.167376 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.168045 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.194042 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.202806 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.218254 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.221243 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.235237 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.258834 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.291406 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.368699 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.444051 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.560932 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.750887 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.751365 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.838324 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.844350 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.952903 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 15 00:13:06 crc kubenswrapper[4861]: I0315 00:13:06.962406 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.092668 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.202254 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.223717 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.234348 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.311982 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.316495 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.508850 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.510984 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.589721 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.663110 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.841504 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 15 00:13:07 crc kubenswrapper[4861]: I0315 00:13:07.922643 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.029365 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.037201 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.086811 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.107030 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.239443 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.275997 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.286528 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.307204 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.401407 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.412609 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.416727 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.447824 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.459308 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.495089 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.653145 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.703996 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.772261 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.864632 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.904503 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 15 00:13:08 crc kubenswrapper[4861]: I0315 00:13:08.915381 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.049794 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.056983 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.088534 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.230188 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.268731 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.310858 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.452353 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.475550 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.522086 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.523302 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.570731 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.641315 4861 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.829285 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.872697 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.902301 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.910172 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 15 00:13:09 crc kubenswrapper[4861]: I0315 00:13:09.970260 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.019338 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.082388 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.104333 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.197646 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.269496 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.314986 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.364653 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.393855 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.504147 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.600203 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.633361 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.691933 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.909754 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.936489 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 15 00:13:10 crc kubenswrapper[4861]: I0315 00:13:10.973364 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.099492 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.141950 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.158768 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.159937 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.165342 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.167230 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.257774 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.325462 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.457083 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.513378 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.536060 4861 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.609618 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.673171 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.701738 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.731913 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.748909 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.830578 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.854233 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.859914 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 15 00:13:11 crc kubenswrapper[4861]: I0315 00:13:11.908060 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.277715 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.300167 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.300186 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.329607 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.350317 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.373977 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.390176 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.529371 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.557154 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.816168 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.885938 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 15 00:13:12 crc kubenswrapper[4861]: I0315 00:13:12.961729 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.089822 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.153125 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.156791 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.175292 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.191965 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.219761 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.356153 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.362763 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.419167 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.498377 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.516380 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.533922 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.668026 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.733510 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.791851 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.810213 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.815090 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.847963 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.861962 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.881981 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 15 00:13:13 crc kubenswrapper[4861]: I0315 00:13:13.901917 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.052821 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.104499 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.147809 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.410822 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.533226 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.629386 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.708156 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.734877 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.766082 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.776442 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.788407 4861 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.809947 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 15 00:13:14 crc kubenswrapper[4861]: I0315 00:13:14.876789 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.027759 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.031496 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.094624 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.143764 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.146190 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.498509 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.501232 4861 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.506718 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-t4ghv" podStartSLOduration=326.506684282 podStartE2EDuration="5m26.506684282s" podCreationTimestamp="2026-03-15 00:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:12:57.65026536 +0000 UTC m=+370.962462933" watchObservedRunningTime="2026-03-15 00:13:15.506684282 +0000 UTC m=+388.818881855" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512036 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j2krj","openshift-kube-apiserver/kube-apiserver-crc"] Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512144 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7f8484fbcc-8scxp","openshift-kube-apiserver/kube-apiserver-crc"] Mar 15 00:13:15 crc kubenswrapper[4861]: E0315 00:13:15.512533 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" containerName="installer" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512607 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" containerName="installer" Mar 15 00:13:15 crc kubenswrapper[4861]: E0315 00:13:15.512627 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512640 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512738 4861 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512778 4861 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c527691d-7c1c-4780-8a7d-ff270650a7eb" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512834 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d821af9-853a-43b8-8a31-c7df6895c19a" containerName="installer" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.512857 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" containerName="oauth-openshift" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.514755 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-t4ghv"] Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.515064 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.520864 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.523791 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.524592 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.524747 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.524992 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.525202 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.525329 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.525397 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.525676 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.526167 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.527612 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.528589 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.528784 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.532370 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.536847 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.546285 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.572500 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.573173 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.573123891 podStartE2EDuration="24.573123891s" podCreationTimestamp="2026-03-15 00:12:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:13:15.553641302 +0000 UTC m=+388.865838855" watchObservedRunningTime="2026-03-15 00:13:15.573123891 +0000 UTC m=+388.885321454" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607462 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607604 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-policies\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607646 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607706 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607862 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607897 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hzvg\" (UniqueName: \"kubernetes.io/projected/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-kube-api-access-7hzvg\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607931 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.607983 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-error\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608043 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-dir\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608101 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-session\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608169 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608221 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608269 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.608343 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-login\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.664295 4861 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.687033 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710123 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710199 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710240 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hzvg\" (UniqueName: \"kubernetes.io/projected/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-kube-api-access-7hzvg\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710282 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-error\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710323 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-dir\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710361 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-session\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710397 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710429 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.710537 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-dir\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711617 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711657 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-login\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711698 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711717 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-policies\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711734 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711753 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-service-ca\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.712374 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.711764 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.712578 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-audit-policies\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.713823 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.719870 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.720041 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.720499 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-error\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.721224 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-session\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.721766 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-login\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.723912 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-router-certs\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.724203 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.724450 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.740448 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hzvg\" (UniqueName: \"kubernetes.io/projected/d32eb072-27d3-4e39-9eda-e5f9eb05f02e-kube-api-access-7hzvg\") pod \"oauth-openshift-7f8484fbcc-8scxp\" (UID: \"d32eb072-27d3-4e39-9eda-e5f9eb05f02e\") " pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.760917 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.787138 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 15 00:13:15 crc kubenswrapper[4861]: I0315 00:13:15.864380 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.013846 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.016339 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.049279 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.197788 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.497080 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.538150 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.749043 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.881168 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 15 00:13:16 crc kubenswrapper[4861]: I0315 00:13:16.927007 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.003174 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f8484fbcc-8scxp"] Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.100912 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.421009 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e8992d1-2a3f-4a3b-b857-821cfb8bed31" path="/var/lib/kubelet/pods/1e8992d1-2a3f-4a3b-b857-821cfb8bed31/volumes" Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.496996 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.581292 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7f8484fbcc-8scxp"] Mar 15 00:13:17 crc kubenswrapper[4861]: W0315 00:13:17.588636 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd32eb072_27d3_4e39_9eda_e5f9eb05f02e.slice/crio-f42e8a34c37d1f9cebd16000e26d52398243c2084f3e482954c1211ad64b421a WatchSource:0}: Error finding container f42e8a34c37d1f9cebd16000e26d52398243c2084f3e482954c1211ad64b421a: Status 404 returned error can't find the container with id f42e8a34c37d1f9cebd16000e26d52398243c2084f3e482954c1211ad64b421a Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.592940 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.774607 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" event={"ID":"d32eb072-27d3-4e39-9eda-e5f9eb05f02e","Type":"ContainerStarted","Data":"f42e8a34c37d1f9cebd16000e26d52398243c2084f3e482954c1211ad64b421a"} Mar 15 00:13:17 crc kubenswrapper[4861]: I0315 00:13:17.786460 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 15 00:13:18 crc kubenswrapper[4861]: I0315 00:13:18.163531 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 15 00:13:18 crc kubenswrapper[4861]: I0315 00:13:18.787107 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" event={"ID":"d32eb072-27d3-4e39-9eda-e5f9eb05f02e","Type":"ContainerStarted","Data":"6f601c16e175eafee342ac84ae1843dff9b6aaec0a6c3531faf82359476efc93"} Mar 15 00:13:18 crc kubenswrapper[4861]: I0315 00:13:18.787806 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:18 crc kubenswrapper[4861]: I0315 00:13:18.803073 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" Mar 15 00:13:18 crc kubenswrapper[4861]: I0315 00:13:18.823250 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7f8484fbcc-8scxp" podStartSLOduration=61.823224444 podStartE2EDuration="1m1.823224444s" podCreationTimestamp="2026-03-15 00:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:13:18.818512488 +0000 UTC m=+392.130710111" watchObservedRunningTime="2026-03-15 00:13:18.823224444 +0000 UTC m=+392.135422027" Mar 15 00:13:24 crc kubenswrapper[4861]: I0315 00:13:24.580119 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 15 00:13:25 crc kubenswrapper[4861]: I0315 00:13:25.528192 4861 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 15 00:13:25 crc kubenswrapper[4861]: I0315 00:13:25.528726 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://eb3b181c483ec061d4b716a4f46ff5bc65b7c1f79a65bf5ef0a8beb64eeb0f76" gracePeriod=5 Mar 15 00:13:30 crc kubenswrapper[4861]: I0315 00:13:30.886076 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 15 00:13:30 crc kubenswrapper[4861]: I0315 00:13:30.886935 4861 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="eb3b181c483ec061d4b716a4f46ff5bc65b7c1f79a65bf5ef0a8beb64eeb0f76" exitCode=137 Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.141769 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.141931 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342198 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342323 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342410 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342404 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342444 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342514 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342536 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342588 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342605 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342941 4861 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.342975 4861 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.343001 4861 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.343023 4861 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.355308 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.422866 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.444687 4861 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.898041 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.898164 4861 scope.go:117] "RemoveContainer" containerID="eb3b181c483ec061d4b716a4f46ff5bc65b7c1f79a65bf5ef0a8beb64eeb0f76" Mar 15 00:13:31 crc kubenswrapper[4861]: I0315 00:13:31.898272 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 15 00:13:33 crc kubenswrapper[4861]: I0315 00:13:33.854287 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 15 00:13:35 crc kubenswrapper[4861]: I0315 00:13:35.190275 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 15 00:13:37 crc kubenswrapper[4861]: I0315 00:13:37.952986 4861 generic.go:334] "Generic (PLEG): container finished" podID="85612e8d-2528-46f1-8436-c738e2961cff" containerID="5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc" exitCode=0 Mar 15 00:13:37 crc kubenswrapper[4861]: I0315 00:13:37.953066 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerDied","Data":"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc"} Mar 15 00:13:37 crc kubenswrapper[4861]: I0315 00:13:37.954004 4861 scope.go:117] "RemoveContainer" containerID="5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc" Mar 15 00:13:38 crc kubenswrapper[4861]: I0315 00:13:38.965282 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerStarted","Data":"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8"} Mar 15 00:13:38 crc kubenswrapper[4861]: I0315 00:13:38.966097 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:13:38 crc kubenswrapper[4861]: I0315 00:13:38.968378 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:13:40 crc kubenswrapper[4861]: I0315 00:13:40.282537 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 15 00:13:40 crc kubenswrapper[4861]: I0315 00:13:40.788075 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 15 00:13:41 crc kubenswrapper[4861]: I0315 00:13:41.789721 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 15 00:13:44 crc kubenswrapper[4861]: I0315 00:13:44.008988 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 15 00:13:44 crc kubenswrapper[4861]: I0315 00:13:44.191139 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 15 00:13:46 crc kubenswrapper[4861]: I0315 00:13:46.499198 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 15 00:13:46 crc kubenswrapper[4861]: I0315 00:13:46.736857 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 15 00:13:47 crc kubenswrapper[4861]: I0315 00:13:47.104031 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 15 00:13:49 crc kubenswrapper[4861]: I0315 00:13:49.653747 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 15 00:13:52 crc kubenswrapper[4861]: I0315 00:13:52.266851 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 15 00:13:53 crc kubenswrapper[4861]: I0315 00:13:53.692901 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 15 00:13:56 crc kubenswrapper[4861]: I0315 00:13:56.826277 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.021256 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.201936 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558894-d7mzm"] Mar 15 00:14:00 crc kubenswrapper[4861]: E0315 00:14:00.202339 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.202368 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.202492 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.203027 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.206957 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.207153 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.224378 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.225646 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558894-d7mzm"] Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.294098 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnskm\" (UniqueName: \"kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm\") pod \"auto-csr-approver-29558894-d7mzm\" (UID: \"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a\") " pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.395214 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnskm\" (UniqueName: \"kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm\") pod \"auto-csr-approver-29558894-d7mzm\" (UID: \"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a\") " pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.431616 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnskm\" (UniqueName: \"kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm\") pod \"auto-csr-approver-29558894-d7mzm\" (UID: \"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a\") " pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.541915 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:00 crc kubenswrapper[4861]: I0315 00:14:00.779382 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558894-d7mzm"] Mar 15 00:14:01 crc kubenswrapper[4861]: I0315 00:14:01.126331 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" event={"ID":"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a","Type":"ContainerStarted","Data":"fe7eba70f56cbbaea0022e8e15430f298045433c4f0e703d36afd78fe1213d57"} Mar 15 00:14:02 crc kubenswrapper[4861]: I0315 00:14:02.133762 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" event={"ID":"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a","Type":"ContainerStarted","Data":"430ae6b9ad7d2ba2b6fdd32abf6b778ffb46c52025c3f506fcb15dfbfbf5523e"} Mar 15 00:14:02 crc kubenswrapper[4861]: I0315 00:14:02.152618 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" podStartSLOduration=1.3387263329999999 podStartE2EDuration="2.152593818s" podCreationTimestamp="2026-03-15 00:14:00 +0000 UTC" firstStartedPulling="2026-03-15 00:14:00.800352046 +0000 UTC m=+434.112549599" lastFinishedPulling="2026-03-15 00:14:01.614219551 +0000 UTC m=+434.926417084" observedRunningTime="2026-03-15 00:14:02.147745219 +0000 UTC m=+435.459942752" watchObservedRunningTime="2026-03-15 00:14:02.152593818 +0000 UTC m=+435.464791351" Mar 15 00:14:02 crc kubenswrapper[4861]: I0315 00:14:02.282159 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:14:02 crc kubenswrapper[4861]: I0315 00:14:02.282240 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:14:03 crc kubenswrapper[4861]: I0315 00:14:03.143256 4861 generic.go:334] "Generic (PLEG): container finished" podID="d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" containerID="430ae6b9ad7d2ba2b6fdd32abf6b778ffb46c52025c3f506fcb15dfbfbf5523e" exitCode=0 Mar 15 00:14:03 crc kubenswrapper[4861]: I0315 00:14:03.143477 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" event={"ID":"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a","Type":"ContainerDied","Data":"430ae6b9ad7d2ba2b6fdd32abf6b778ffb46c52025c3f506fcb15dfbfbf5523e"} Mar 15 00:14:03 crc kubenswrapper[4861]: I0315 00:14:03.148156 4861 generic.go:334] "Generic (PLEG): container finished" podID="981ba86e-1a94-4acc-ae13-25da019d91fd" containerID="006be2d24ad8ed61eec2ba7fd726728425f2f04791bcf16bc7071441bd028fbc" exitCode=0 Mar 15 00:14:03 crc kubenswrapper[4861]: I0315 00:14:03.148199 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29558880-xw69x" event={"ID":"981ba86e-1a94-4acc-ae13-25da019d91fd","Type":"ContainerDied","Data":"006be2d24ad8ed61eec2ba7fd726728425f2f04791bcf16bc7071441bd028fbc"} Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.544094 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.548531 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.659464 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca\") pod \"981ba86e-1a94-4acc-ae13-25da019d91fd\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.659606 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bkkl\" (UniqueName: \"kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl\") pod \"981ba86e-1a94-4acc-ae13-25da019d91fd\" (UID: \"981ba86e-1a94-4acc-ae13-25da019d91fd\") " Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.659867 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnskm\" (UniqueName: \"kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm\") pod \"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a\" (UID: \"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a\") " Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.661093 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca" (OuterVolumeSpecName: "serviceca") pod "981ba86e-1a94-4acc-ae13-25da019d91fd" (UID: "981ba86e-1a94-4acc-ae13-25da019d91fd"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.669847 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm" (OuterVolumeSpecName: "kube-api-access-hnskm") pod "d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" (UID: "d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a"). InnerVolumeSpecName "kube-api-access-hnskm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.670391 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl" (OuterVolumeSpecName: "kube-api-access-9bkkl") pod "981ba86e-1a94-4acc-ae13-25da019d91fd" (UID: "981ba86e-1a94-4acc-ae13-25da019d91fd"). InnerVolumeSpecName "kube-api-access-9bkkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.761611 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnskm\" (UniqueName: \"kubernetes.io/projected/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a-kube-api-access-hnskm\") on node \"crc\" DevicePath \"\"" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.761657 4861 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/981ba86e-1a94-4acc-ae13-25da019d91fd-serviceca\") on node \"crc\" DevicePath \"\"" Mar 15 00:14:04 crc kubenswrapper[4861]: I0315 00:14:04.761669 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bkkl\" (UniqueName: \"kubernetes.io/projected/981ba86e-1a94-4acc-ae13-25da019d91fd-kube-api-access-9bkkl\") on node \"crc\" DevicePath \"\"" Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.166066 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" event={"ID":"d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a","Type":"ContainerDied","Data":"fe7eba70f56cbbaea0022e8e15430f298045433c4f0e703d36afd78fe1213d57"} Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.166638 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe7eba70f56cbbaea0022e8e15430f298045433c4f0e703d36afd78fe1213d57" Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.166468 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558894-d7mzm" Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.172161 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29558880-xw69x" event={"ID":"981ba86e-1a94-4acc-ae13-25da019d91fd","Type":"ContainerDied","Data":"d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da"} Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.172220 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d12a30d9e9e93c55523af0262bd023f019e3defa387d03b2db39b7cbd44222da" Mar 15 00:14:05 crc kubenswrapper[4861]: I0315 00:14:05.172261 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29558880-xw69x" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.848450 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmwmz"] Mar 15 00:14:28 crc kubenswrapper[4861]: E0315 00:14:28.849595 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" containerName="oc" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.849618 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" containerName="oc" Mar 15 00:14:28 crc kubenswrapper[4861]: E0315 00:14:28.849645 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="981ba86e-1a94-4acc-ae13-25da019d91fd" containerName="image-pruner" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.849653 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="981ba86e-1a94-4acc-ae13-25da019d91fd" containerName="image-pruner" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.849790 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" containerName="oc" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.849821 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="981ba86e-1a94-4acc-ae13-25da019d91fd" containerName="image-pruner" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.850319 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:28 crc kubenswrapper[4861]: I0315 00:14:28.862342 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmwmz"] Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.025955 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026464 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bcb28a88-158c-4e29-b93e-f654f3895e47-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026499 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-tls\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026531 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-trusted-ca\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026695 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bcb28a88-158c-4e29-b93e-f654f3895e47-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026781 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxkd4\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-kube-api-access-gxkd4\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026807 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-certificates\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.026854 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-bound-sa-token\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.065003 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128784 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-trusted-ca\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128857 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bcb28a88-158c-4e29-b93e-f654f3895e47-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128887 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-certificates\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128904 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxkd4\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-kube-api-access-gxkd4\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128924 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-bound-sa-token\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128961 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bcb28a88-158c-4e29-b93e-f654f3895e47-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.128989 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-tls\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.129540 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bcb28a88-158c-4e29-b93e-f654f3895e47-ca-trust-extracted\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.130287 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-trusted-ca\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.132528 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-certificates\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.138688 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bcb28a88-158c-4e29-b93e-f654f3895e47-installation-pull-secrets\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.142546 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-registry-tls\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.146874 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxkd4\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-kube-api-access-gxkd4\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.147156 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bcb28a88-158c-4e29-b93e-f654f3895e47-bound-sa-token\") pod \"image-registry-66df7c8f76-fmwmz\" (UID: \"bcb28a88-158c-4e29-b93e-f654f3895e47\") " pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:29 crc kubenswrapper[4861]: I0315 00:14:29.169109 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:30 crc kubenswrapper[4861]: I0315 00:14:30.078770 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-fmwmz"] Mar 15 00:14:30 crc kubenswrapper[4861]: I0315 00:14:30.364574 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" event={"ID":"bcb28a88-158c-4e29-b93e-f654f3895e47","Type":"ContainerStarted","Data":"8663328db135acde3c69b390b6acdebf7d001d7a6e1ff864e2a5d1f399a163c0"} Mar 15 00:14:30 crc kubenswrapper[4861]: I0315 00:14:30.365068 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" event={"ID":"bcb28a88-158c-4e29-b93e-f654f3895e47","Type":"ContainerStarted","Data":"c5183193ffafe9fe026d631e5effbab4de0bdb0f62cfa30b7e5464fa3a6d0a50"} Mar 15 00:14:30 crc kubenswrapper[4861]: I0315 00:14:30.365100 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:30 crc kubenswrapper[4861]: I0315 00:14:30.387348 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" podStartSLOduration=2.387327012 podStartE2EDuration="2.387327012s" podCreationTimestamp="2026-03-15 00:14:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:14:30.385446991 +0000 UTC m=+463.697644524" watchObservedRunningTime="2026-03-15 00:14:30.387327012 +0000 UTC m=+463.699524545" Mar 15 00:14:32 crc kubenswrapper[4861]: I0315 00:14:32.282164 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:14:32 crc kubenswrapper[4861]: I0315 00:14:32.282780 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:14:49 crc kubenswrapper[4861]: I0315 00:14:49.175193 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-fmwmz" Mar 15 00:14:49 crc kubenswrapper[4861]: I0315 00:14:49.247988 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.168292 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws"] Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.173194 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.179429 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.179698 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.188180 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws"] Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.349246 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwl8\" (UniqueName: \"kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.349358 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.349425 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.451393 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwl8\" (UniqueName: \"kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.451469 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.451586 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.453210 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.467225 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.489869 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwl8\" (UniqueName: \"kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8\") pod \"collect-profiles-29558895-vdgws\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.511293 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:00 crc kubenswrapper[4861]: I0315 00:15:00.808249 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws"] Mar 15 00:15:01 crc kubenswrapper[4861]: I0315 00:15:01.590502 4861 generic.go:334] "Generic (PLEG): container finished" podID="bc769003-a58a-462d-b631-48aa61e52a20" containerID="5dd14899dec198c6ead693aa47f2c07f1dc86139f6471adbf90f10968cd8373b" exitCode=0 Mar 15 00:15:01 crc kubenswrapper[4861]: I0315 00:15:01.590626 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" event={"ID":"bc769003-a58a-462d-b631-48aa61e52a20","Type":"ContainerDied","Data":"5dd14899dec198c6ead693aa47f2c07f1dc86139f6471adbf90f10968cd8373b"} Mar 15 00:15:01 crc kubenswrapper[4861]: I0315 00:15:01.591124 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" event={"ID":"bc769003-a58a-462d-b631-48aa61e52a20","Type":"ContainerStarted","Data":"4b127241970c12eb83ebb6eaa39d67915fb6b70eee4f2360ec6db3d86acb85f1"} Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.282502 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.282606 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.282684 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.283346 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.283402 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db" gracePeriod=600 Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.606065 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db" exitCode=0 Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.606175 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db"} Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.606933 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e"} Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.606992 4861 scope.go:117] "RemoveContainer" containerID="e329f5ddf4a151bb5bed759c93083ab6b9477d85028327861cb4762084acd390" Mar 15 00:15:02 crc kubenswrapper[4861]: I0315 00:15:02.951630 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.118143 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume\") pod \"bc769003-a58a-462d-b631-48aa61e52a20\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.118491 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume\") pod \"bc769003-a58a-462d-b631-48aa61e52a20\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.118607 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvwl8\" (UniqueName: \"kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8\") pod \"bc769003-a58a-462d-b631-48aa61e52a20\" (UID: \"bc769003-a58a-462d-b631-48aa61e52a20\") " Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.119576 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume" (OuterVolumeSpecName: "config-volume") pod "bc769003-a58a-462d-b631-48aa61e52a20" (UID: "bc769003-a58a-462d-b631-48aa61e52a20"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.126632 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8" (OuterVolumeSpecName: "kube-api-access-mvwl8") pod "bc769003-a58a-462d-b631-48aa61e52a20" (UID: "bc769003-a58a-462d-b631-48aa61e52a20"). InnerVolumeSpecName "kube-api-access-mvwl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.127866 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bc769003-a58a-462d-b631-48aa61e52a20" (UID: "bc769003-a58a-462d-b631-48aa61e52a20"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.219996 4861 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bc769003-a58a-462d-b631-48aa61e52a20-config-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.220048 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvwl8\" (UniqueName: \"kubernetes.io/projected/bc769003-a58a-462d-b631-48aa61e52a20-kube-api-access-mvwl8\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.220070 4861 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bc769003-a58a-462d-b631-48aa61e52a20-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.622427 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" event={"ID":"bc769003-a58a-462d-b631-48aa61e52a20","Type":"ContainerDied","Data":"4b127241970c12eb83ebb6eaa39d67915fb6b70eee4f2360ec6db3d86acb85f1"} Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.622503 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b127241970c12eb83ebb6eaa39d67915fb6b70eee4f2360ec6db3d86acb85f1" Mar 15 00:15:03 crc kubenswrapper[4861]: I0315 00:15:03.622502 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558895-vdgws" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.166409 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.167413 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9rrjp" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="registry-server" containerID="cri-o://b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286" gracePeriod=30 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.176712 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.177056 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zz6cr" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="registry-server" containerID="cri-o://fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245" gracePeriod=30 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.199398 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.199865 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" containerID="cri-o://8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8" gracePeriod=30 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.215101 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.215837 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2jfx8" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="registry-server" containerID="cri-o://32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab" gracePeriod=30 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.221456 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-smdks"] Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.221744 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc769003-a58a-462d-b631-48aa61e52a20" containerName="collect-profiles" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.221764 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc769003-a58a-462d-b631-48aa61e52a20" containerName="collect-profiles" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.221896 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc769003-a58a-462d-b631-48aa61e52a20" containerName="collect-profiles" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.222370 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.229110 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.229484 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9447z" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="registry-server" containerID="cri-o://69e3a76c57af1a427485c1b19a18e50fc576f41ba430b2cb8cc0b770cfea5966" gracePeriod=30 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.240598 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-smdks"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.332893 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.332996 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.333103 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdrl9\" (UniqueName: \"kubernetes.io/projected/d7fbc705-3f2f-47d2-9b12-c308b061a5db-kube-api-access-zdrl9\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.434672 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.434727 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.434762 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdrl9\" (UniqueName: \"kubernetes.io/projected/d7fbc705-3f2f-47d2-9b12-c308b061a5db-kube-api-access-zdrl9\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.436304 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.449571 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d7fbc705-3f2f-47d2-9b12-c308b061a5db-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.455075 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdrl9\" (UniqueName: \"kubernetes.io/projected/d7fbc705-3f2f-47d2-9b12-c308b061a5db-kube-api-access-zdrl9\") pod \"marketplace-operator-79b997595-smdks\" (UID: \"d7fbc705-3f2f-47d2-9b12-c308b061a5db\") " pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.551008 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.615370 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.620274 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.625195 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637331 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637359 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics\") pod \"85612e8d-2528-46f1-8436-c738e2961cff\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637406 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhdmw\" (UniqueName: \"kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw\") pod \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637442 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities\") pod \"f3e7027d-525d-48f6-aa18-bb6a622331e6\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637466 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content\") pod \"f3e7027d-525d-48f6-aa18-bb6a622331e6\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637546 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8h5m\" (UniqueName: \"kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m\") pod \"85612e8d-2528-46f1-8436-c738e2961cff\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637590 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities\") pod \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637617 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca\") pod \"85612e8d-2528-46f1-8436-c738e2961cff\" (UID: \"85612e8d-2528-46f1-8436-c738e2961cff\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637685 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content\") pod \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\" (UID: \"eaa5b8bf-382f-43a4-a33c-c1002254f2fe\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.637725 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnhzh\" (UniqueName: \"kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh\") pod \"f3e7027d-525d-48f6-aa18-bb6a622331e6\" (UID: \"f3e7027d-525d-48f6-aa18-bb6a622331e6\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.641450 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities" (OuterVolumeSpecName: "utilities") pod "f3e7027d-525d-48f6-aa18-bb6a622331e6" (UID: "f3e7027d-525d-48f6-aa18-bb6a622331e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.641443 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "85612e8d-2528-46f1-8436-c738e2961cff" (UID: "85612e8d-2528-46f1-8436-c738e2961cff"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.642449 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities" (OuterVolumeSpecName: "utilities") pod "eaa5b8bf-382f-43a4-a33c-c1002254f2fe" (UID: "eaa5b8bf-382f-43a4-a33c-c1002254f2fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.644624 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m" (OuterVolumeSpecName: "kube-api-access-d8h5m") pod "85612e8d-2528-46f1-8436-c738e2961cff" (UID: "85612e8d-2528-46f1-8436-c738e2961cff"). InnerVolumeSpecName "kube-api-access-d8h5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.649057 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw" (OuterVolumeSpecName: "kube-api-access-zhdmw") pod "eaa5b8bf-382f-43a4-a33c-c1002254f2fe" (UID: "eaa5b8bf-382f-43a4-a33c-c1002254f2fe"). InnerVolumeSpecName "kube-api-access-zhdmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.649273 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh" (OuterVolumeSpecName: "kube-api-access-hnhzh") pod "f3e7027d-525d-48f6-aa18-bb6a622331e6" (UID: "f3e7027d-525d-48f6-aa18-bb6a622331e6"). InnerVolumeSpecName "kube-api-access-hnhzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.671361 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "85612e8d-2528-46f1-8436-c738e2961cff" (UID: "85612e8d-2528-46f1-8436-c738e2961cff"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.693334 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerID="32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab" exitCode=0 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.693431 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerDied","Data":"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.693470 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2jfx8" event={"ID":"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1","Type":"ContainerDied","Data":"e77745dd848943e2cf9ae52843c30f6cb6f286373e828d58bf47caf852e5a10c"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.693492 4861 scope.go:117] "RemoveContainer" containerID="32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.693891 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2jfx8" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.705092 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.712985 4861 generic.go:334] "Generic (PLEG): container finished" podID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerID="fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245" exitCode=0 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.713055 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerDied","Data":"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.713093 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6cr" event={"ID":"f3e7027d-525d-48f6-aa18-bb6a622331e6","Type":"ContainerDied","Data":"ff68155b833de13ab267451281245c88c49206e8c1fe79fe87d771773d515c51"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.713196 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6cr" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.738159 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities\") pod \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741132 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities\") pod \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741238 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content\") pod \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741355 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content\") pod \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741469 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd8xn\" (UniqueName: \"kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn\") pod \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\" (UID: \"ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741540 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9f56\" (UniqueName: \"kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56\") pod \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\" (UID: \"0c5fd4bf-9989-488e-83e3-a41e0221c18e\") " Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741812 4861 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741870 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhdmw\" (UniqueName: \"kubernetes.io/projected/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-kube-api-access-zhdmw\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741930 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.741979 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8h5m\" (UniqueName: \"kubernetes.io/projected/85612e8d-2528-46f1-8436-c738e2961cff-kube-api-access-d8h5m\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.742027 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.742088 4861 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85612e8d-2528-46f1-8436-c738e2961cff-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.742189 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnhzh\" (UniqueName: \"kubernetes.io/projected/f3e7027d-525d-48f6-aa18-bb6a622331e6-kube-api-access-hnhzh\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.745955 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eaa5b8bf-382f-43a4-a33c-c1002254f2fe" (UID: "eaa5b8bf-382f-43a4-a33c-c1002254f2fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.749608 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities" (OuterVolumeSpecName: "utilities") pod "0c5fd4bf-9989-488e-83e3-a41e0221c18e" (UID: "0c5fd4bf-9989-488e-83e3-a41e0221c18e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.749762 4861 generic.go:334] "Generic (PLEG): container finished" podID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerID="b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286" exitCode=0 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.749831 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerDied","Data":"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.749872 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9rrjp" event={"ID":"eaa5b8bf-382f-43a4-a33c-c1002254f2fe","Type":"ContainerDied","Data":"74fe1e3996101a89ab628a00d004f4a6b6576500c18cd575b5e58fc70458920d"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.749952 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9rrjp" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.758475 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3e7027d-525d-48f6-aa18-bb6a622331e6" (UID: "f3e7027d-525d-48f6-aa18-bb6a622331e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.770802 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn" (OuterVolumeSpecName: "kube-api-access-fd8xn") pod "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" (UID: "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1"). InnerVolumeSpecName "kube-api-access-fd8xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.770854 4861 scope.go:117] "RemoveContainer" containerID="f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.773495 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities" (OuterVolumeSpecName: "utilities") pod "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" (UID: "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.773859 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56" (OuterVolumeSpecName: "kube-api-access-j9f56") pod "0c5fd4bf-9989-488e-83e3-a41e0221c18e" (UID: "0c5fd4bf-9989-488e-83e3-a41e0221c18e"). InnerVolumeSpecName "kube-api-access-j9f56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.775831 4861 generic.go:334] "Generic (PLEG): container finished" podID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerID="69e3a76c57af1a427485c1b19a18e50fc576f41ba430b2cb8cc0b770cfea5966" exitCode=0 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.776492 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9447z" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.776634 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9447z" event={"ID":"0c5fd4bf-9989-488e-83e3-a41e0221c18e","Type":"ContainerDied","Data":"69e3a76c57af1a427485c1b19a18e50fc576f41ba430b2cb8cc0b770cfea5966"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.778458 4861 generic.go:334] "Generic (PLEG): container finished" podID="85612e8d-2528-46f1-8436-c738e2961cff" containerID="8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8" exitCode=0 Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.778499 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerDied","Data":"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.778527 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" event={"ID":"85612e8d-2528-46f1-8436-c738e2961cff","Type":"ContainerDied","Data":"7fa5d6ee90d0d1e86542acbda0ec6ef580bc847d471f1e06e1ddeee829b2acf9"} Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.778584 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kt6gz" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.789436 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.792995 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9rrjp"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.808190 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" (UID: "ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.808376 4861 scope.go:117] "RemoveContainer" containerID="6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.825409 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.839997 4861 scope.go:117] "RemoveContainer" containerID="32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.841078 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab\": container with ID starting with 32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab not found: ID does not exist" containerID="32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.841127 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab"} err="failed to get container status \"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab\": rpc error: code = NotFound desc = could not find container \"32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab\": container with ID starting with 32e458458fe4183148bb136c780020ddeca40f287530655a204291c71b24cdab not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.841156 4861 scope.go:117] "RemoveContainer" containerID="f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.841532 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98\": container with ID starting with f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98 not found: ID does not exist" containerID="f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.841638 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98"} err="failed to get container status \"f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98\": rpc error: code = NotFound desc = could not find container \"f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98\": container with ID starting with f93384c1069256f454abe4aa3098b5da55e7f4e6b764c75c68ea19156f138f98 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.841678 4861 scope.go:117] "RemoveContainer" containerID="6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.841959 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18\": container with ID starting with 6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18 not found: ID does not exist" containerID="6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.841986 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18"} err="failed to get container status \"6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18\": rpc error: code = NotFound desc = could not find container \"6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18\": container with ID starting with 6f1a4cb40e0541c11f627cc7a8a12e30bc5dade7605d329a41b3df6ab7fa9d18 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842018 4861 scope.go:117] "RemoveContainer" containerID="fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842869 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3e7027d-525d-48f6-aa18-bb6a622331e6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842898 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd8xn\" (UniqueName: \"kubernetes.io/projected/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-kube-api-access-fd8xn\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842913 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9f56\" (UniqueName: \"kubernetes.io/projected/0c5fd4bf-9989-488e-83e3-a41e0221c18e-kube-api-access-j9f56\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842923 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842937 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842947 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.842957 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eaa5b8bf-382f-43a4-a33c-c1002254f2fe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.859728 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kt6gz"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.859863 4861 scope.go:117] "RemoveContainer" containerID="b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.868852 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-smdks"] Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.878029 4861 scope.go:117] "RemoveContainer" containerID="d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.893075 4861 scope.go:117] "RemoveContainer" containerID="fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.893893 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245\": container with ID starting with fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245 not found: ID does not exist" containerID="fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.893946 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245"} err="failed to get container status \"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245\": rpc error: code = NotFound desc = could not find container \"fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245\": container with ID starting with fd46372d0cdae8c4d965a983c4db949d547f260f09eba8346eb54417926dc245 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.893984 4861 scope.go:117] "RemoveContainer" containerID="b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.894289 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998\": container with ID starting with b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998 not found: ID does not exist" containerID="b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.894312 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998"} err="failed to get container status \"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998\": rpc error: code = NotFound desc = could not find container \"b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998\": container with ID starting with b1484c08f5c2a55f53187875f490de72c10051fb9b0a87df091ce87f11e66998 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.894327 4861 scope.go:117] "RemoveContainer" containerID="d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.894761 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945\": container with ID starting with d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945 not found: ID does not exist" containerID="d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.894794 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945"} err="failed to get container status \"d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945\": rpc error: code = NotFound desc = could not find container \"d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945\": container with ID starting with d99960ab9816c20cea27e4d6f86b240e6d3d98553159528f93ecbc0a39c6e945 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.894810 4861 scope.go:117] "RemoveContainer" containerID="b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.909920 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c5fd4bf-9989-488e-83e3-a41e0221c18e" (UID: "0c5fd4bf-9989-488e-83e3-a41e0221c18e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.918481 4861 scope.go:117] "RemoveContainer" containerID="6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.951520 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5fd4bf-9989-488e-83e3-a41e0221c18e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.969202 4861 scope.go:117] "RemoveContainer" containerID="315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.990591 4861 scope.go:117] "RemoveContainer" containerID="b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.991522 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286\": container with ID starting with b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286 not found: ID does not exist" containerID="b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.991600 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286"} err="failed to get container status \"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286\": rpc error: code = NotFound desc = could not find container \"b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286\": container with ID starting with b37fd2e2dfd3b6550dbb4f398f3b15aba5bebaab132af28dc0e8f0656dd3c286 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.991641 4861 scope.go:117] "RemoveContainer" containerID="6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.992067 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305\": container with ID starting with 6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305 not found: ID does not exist" containerID="6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.992092 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305"} err="failed to get container status \"6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305\": rpc error: code = NotFound desc = could not find container \"6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305\": container with ID starting with 6339cf5d835ae9fd62d72c49816e1757e27d6238fa870bd1f18884f0bce5b305 not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.992108 4861 scope.go:117] "RemoveContainer" containerID="315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f" Mar 15 00:15:09 crc kubenswrapper[4861]: E0315 00:15:09.992389 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f\": container with ID starting with 315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f not found: ID does not exist" containerID="315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.992411 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f"} err="failed to get container status \"315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f\": rpc error: code = NotFound desc = could not find container \"315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f\": container with ID starting with 315b9d8e88a5b0a85a350079e04c5a7aa5d6c0a986d9dc4484e41ce64ed6aa6f not found: ID does not exist" Mar 15 00:15:09 crc kubenswrapper[4861]: I0315 00:15:09.992426 4861 scope.go:117] "RemoveContainer" containerID="69e3a76c57af1a427485c1b19a18e50fc576f41ba430b2cb8cc0b770cfea5966" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.010325 4861 scope.go:117] "RemoveContainer" containerID="21235c8b1e36b19a097af2b256db464f1d71853780199e8fcdbf8d5f87cc9c8b" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.029129 4861 scope.go:117] "RemoveContainer" containerID="cb04a90340acb6cf4df240657fb8099c22b2d6589e2d2c3703a0e616c53b36ed" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.057014 4861 scope.go:117] "RemoveContainer" containerID="8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.057238 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.060127 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zz6cr"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.074634 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.078343 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2jfx8"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.079188 4861 scope.go:117] "RemoveContainer" containerID="5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.102382 4861 scope.go:117] "RemoveContainer" containerID="8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8" Mar 15 00:15:10 crc kubenswrapper[4861]: E0315 00:15:10.103202 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8\": container with ID starting with 8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8 not found: ID does not exist" containerID="8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.103250 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8"} err="failed to get container status \"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8\": rpc error: code = NotFound desc = could not find container \"8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8\": container with ID starting with 8df2990bc21da9ff3ee54a3c122ac1e7a64d11af3ca261c51c0135b3af6ecef8 not found: ID does not exist" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.103286 4861 scope.go:117] "RemoveContainer" containerID="5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc" Mar 15 00:15:10 crc kubenswrapper[4861]: E0315 00:15:10.103839 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc\": container with ID starting with 5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc not found: ID does not exist" containerID="5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.103872 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc"} err="failed to get container status \"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc\": rpc error: code = NotFound desc = could not find container \"5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc\": container with ID starting with 5467f550a05d5802eaea5ef476f63ad5bf9353e44c7e102b3d348f1eb5bbe0dc not found: ID does not exist" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.105683 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.109995 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9447z"] Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.815854 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" event={"ID":"d7fbc705-3f2f-47d2-9b12-c308b061a5db","Type":"ContainerStarted","Data":"cd7b5f04758e14278c7f2c38627abd4d012900d962d23bd179a3664637018476"} Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.816355 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.816378 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" event={"ID":"d7fbc705-3f2f-47d2-9b12-c308b061a5db","Type":"ContainerStarted","Data":"93c89e1e1a7b69f8e89facd1377baef14a9a5c7b01440d3a97a803c3ce2c1066"} Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.825067 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" Mar 15 00:15:10 crc kubenswrapper[4861]: I0315 00:15:10.839071 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-smdks" podStartSLOduration=1.839045702 podStartE2EDuration="1.839045702s" podCreationTimestamp="2026-03-15 00:15:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:15:10.838054426 +0000 UTC m=+504.150251969" watchObservedRunningTime="2026-03-15 00:15:10.839045702 +0000 UTC m=+504.151243245" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.393822 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394191 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394214 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394234 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394250 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394265 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394279 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394328 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394342 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394357 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394369 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394392 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394406 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394420 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394431 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394447 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394460 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394479 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394490 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394528 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394541 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394584 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394600 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394618 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394630 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="extract-content" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.394649 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394661 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="extract-utilities" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394832 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394861 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394876 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394892 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394910 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.394934 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" containerName="registry-server" Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.395082 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.395095 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="85612e8d-2528-46f1-8436-c738e2961cff" containerName="marketplace-operator" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.397388 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.401948 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.420019 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5fd4bf-9989-488e-83e3-a41e0221c18e" path="/var/lib/kubelet/pods/0c5fd4bf-9989-488e-83e3-a41e0221c18e/volumes" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.420746 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85612e8d-2528-46f1-8436-c738e2961cff" path="/var/lib/kubelet/pods/85612e8d-2528-46f1-8436-c738e2961cff/volumes" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.421302 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1" path="/var/lib/kubelet/pods/ce66aba6-9ffe-44d9-9c90-961ab7d5a4e1/volumes" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.422715 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa5b8bf-382f-43a4-a33c-c1002254f2fe" path="/var/lib/kubelet/pods/eaa5b8bf-382f-43a4-a33c-c1002254f2fe/volumes" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.423298 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3e7027d-525d-48f6-aa18-bb6a622331e6" path="/var/lib/kubelet/pods/f3e7027d-525d-48f6-aa18-bb6a622331e6/volumes" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.426800 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.472232 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.472342 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt6ch\" (UniqueName: \"kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.472415 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.574031 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.574337 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt6ch\" (UniqueName: \"kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.574491 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.575061 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.575147 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.588014 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w995g"] Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.589460 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: W0315 00:15:11.591315 4861 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 15 00:15:11 crc kubenswrapper[4861]: E0315 00:15:11.591371 4861 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.607089 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt6ch\" (UniqueName: \"kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch\") pod \"redhat-marketplace-ccgz7\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.616942 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w995g"] Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.676989 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-utilities\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.677049 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-catalog-content\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.677095 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zm6c\" (UniqueName: \"kubernetes.io/projected/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-kube-api-access-5zm6c\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.723454 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.778426 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zm6c\" (UniqueName: \"kubernetes.io/projected/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-kube-api-access-5zm6c\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.778529 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-utilities\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.778549 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-catalog-content\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.779310 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-utilities\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.779343 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-catalog-content\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.798595 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zm6c\" (UniqueName: \"kubernetes.io/projected/09972fc0-e7fc-4cd4-87d0-50d858ed9e51-kube-api-access-5zm6c\") pod \"redhat-operators-w995g\" (UID: \"09972fc0-e7fc-4cd4-87d0-50d858ed9e51\") " pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:11 crc kubenswrapper[4861]: I0315 00:15:11.980864 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:15:11 crc kubenswrapper[4861]: W0315 00:15:11.988842 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fb39d55_b4c1_4baf_9d0b_efce980b11c9.slice/crio-3d3f309dce4b8a9883b1c41b8260be3ba51b1f0c08d03ae65cb54abe3a343370 WatchSource:0}: Error finding container 3d3f309dce4b8a9883b1c41b8260be3ba51b1f0c08d03ae65cb54abe3a343370: Status 404 returned error can't find the container with id 3d3f309dce4b8a9883b1c41b8260be3ba51b1f0c08d03ae65cb54abe3a343370 Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.840792 4861 generic.go:334] "Generic (PLEG): container finished" podID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerID="b20434563ff42a55c696bc28ab446ea819347465647c44425fe1043564a8b56f" exitCode=0 Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.840880 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerDied","Data":"b20434563ff42a55c696bc28ab446ea819347465647c44425fe1043564a8b56f"} Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.841363 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerStarted","Data":"3d3f309dce4b8a9883b1c41b8260be3ba51b1f0c08d03ae65cb54abe3a343370"} Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.904955 4861 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/redhat-operators-w995g" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.905636 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:12 crc kubenswrapper[4861]: I0315 00:15:12.931097 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.136966 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w995g"] Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.789019 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pzwlb"] Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.790508 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.792957 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.803348 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzwlb"] Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.838608 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-utilities\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.838941 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-catalog-content\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.839130 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6vvz\" (UniqueName: \"kubernetes.io/projected/47cc9481-9f7a-47f2-933a-93bb592e1bdf-kube-api-access-l6vvz\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.862977 4861 generic.go:334] "Generic (PLEG): container finished" podID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerID="d10f3bbaf7843210c8b1dd84f94ce86c1ef463868de551df4cbbf3d378a80ec0" exitCode=0 Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.863088 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerDied","Data":"d10f3bbaf7843210c8b1dd84f94ce86c1ef463868de551df4cbbf3d378a80ec0"} Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.865328 4861 generic.go:334] "Generic (PLEG): container finished" podID="09972fc0-e7fc-4cd4-87d0-50d858ed9e51" containerID="d41d89db8edafe9d9dfcf945dd4ba9116bbe5cc69d748766ada42d4bcb0b46b9" exitCode=0 Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.865379 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w995g" event={"ID":"09972fc0-e7fc-4cd4-87d0-50d858ed9e51","Type":"ContainerDied","Data":"d41d89db8edafe9d9dfcf945dd4ba9116bbe5cc69d748766ada42d4bcb0b46b9"} Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.865408 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w995g" event={"ID":"09972fc0-e7fc-4cd4-87d0-50d858ed9e51","Type":"ContainerStarted","Data":"9b4ea04cd9834323e29e0038dd3be76dd89c0cead323d32a4caa00df0cc2b223"} Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.940377 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6vvz\" (UniqueName: \"kubernetes.io/projected/47cc9481-9f7a-47f2-933a-93bb592e1bdf-kube-api-access-l6vvz\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.940493 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-utilities\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.940522 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-catalog-content\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.941125 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-catalog-content\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.941306 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47cc9481-9f7a-47f2-933a-93bb592e1bdf-utilities\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.964359 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6vvz\" (UniqueName: \"kubernetes.io/projected/47cc9481-9f7a-47f2-933a-93bb592e1bdf-kube-api-access-l6vvz\") pod \"certified-operators-pzwlb\" (UID: \"47cc9481-9f7a-47f2-933a-93bb592e1bdf\") " pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.994286 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qzm9v"] Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.996358 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:13 crc kubenswrapper[4861]: I0315 00:15:13.998576 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.015924 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzm9v"] Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.042234 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-catalog-content\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.042440 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfsfm\" (UniqueName: \"kubernetes.io/projected/f5a24ae0-68ec-4e10-9592-fa33d3380453-kube-api-access-pfsfm\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.042520 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-utilities\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.144008 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-catalog-content\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.144185 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfsfm\" (UniqueName: \"kubernetes.io/projected/f5a24ae0-68ec-4e10-9592-fa33d3380453-kube-api-access-pfsfm\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.144862 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-utilities\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.144969 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-catalog-content\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.145367 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a24ae0-68ec-4e10-9592-fa33d3380453-utilities\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.177765 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfsfm\" (UniqueName: \"kubernetes.io/projected/f5a24ae0-68ec-4e10-9592-fa33d3380453-kube-api-access-pfsfm\") pod \"community-operators-qzm9v\" (UID: \"f5a24ae0-68ec-4e10-9592-fa33d3380453\") " pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.184162 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.318267 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.320352 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" podUID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" containerName="registry" containerID="cri-o://fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea" gracePeriod=30 Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.427847 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzwlb"] Mar 15 00:15:14 crc kubenswrapper[4861]: W0315 00:15:14.459127 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47cc9481_9f7a_47f2_933a_93bb592e1bdf.slice/crio-7cc8a0cb2ec49afc9a747587d2957fed9c2c97a5966c774f9108efa2bbed8887 WatchSource:0}: Error finding container 7cc8a0cb2ec49afc9a747587d2957fed9c2c97a5966c774f9108efa2bbed8887: Status 404 returned error can't find the container with id 7cc8a0cb2ec49afc9a747587d2957fed9c2c97a5966c774f9108efa2bbed8887 Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.771504 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.819045 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzm9v"] Mar 15 00:15:14 crc kubenswrapper[4861]: W0315 00:15:14.821162 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5a24ae0_68ec_4e10_9592_fa33d3380453.slice/crio-07f109a596df12e3fe90b93c965ddd6eab17b39beaa591390d7c0253d2f1395b WatchSource:0}: Error finding container 07f109a596df12e3fe90b93c965ddd6eab17b39beaa591390d7c0253d2f1395b: Status 404 returned error can't find the container with id 07f109a596df12e3fe90b93c965ddd6eab17b39beaa591390d7c0253d2f1395b Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.855768 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856000 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856057 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856152 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856184 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856223 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856242 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.856313 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhr8f\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f\") pod \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\" (UID: \"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f\") " Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.857766 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.858098 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.866231 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.869100 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.873940 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.876219 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f" (OuterVolumeSpecName: "kube-api-access-xhr8f") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "kube-api-access-xhr8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.877075 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.880065 4861 generic.go:334] "Generic (PLEG): container finished" podID="47cc9481-9f7a-47f2-933a-93bb592e1bdf" containerID="2d79b4326929b7489589368cdeeff7d38d49f93e286c3d795bb76151179988fb" exitCode=0 Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.881208 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzwlb" event={"ID":"47cc9481-9f7a-47f2-933a-93bb592e1bdf","Type":"ContainerDied","Data":"2d79b4326929b7489589368cdeeff7d38d49f93e286c3d795bb76151179988fb"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.881256 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzwlb" event={"ID":"47cc9481-9f7a-47f2-933a-93bb592e1bdf","Type":"ContainerStarted","Data":"7cc8a0cb2ec49afc9a747587d2957fed9c2c97a5966c774f9108efa2bbed8887"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.887234 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerStarted","Data":"f0f3c7320fdb655e7c54f68b728f4f820e835f6801d1f93f7b69b01240ea5f31"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.888897 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzm9v" event={"ID":"f5a24ae0-68ec-4e10-9592-fa33d3380453","Type":"ContainerStarted","Data":"07f109a596df12e3fe90b93c965ddd6eab17b39beaa591390d7c0253d2f1395b"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.893750 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" (UID: "a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.898859 4861 generic.go:334] "Generic (PLEG): container finished" podID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" containerID="fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea" exitCode=0 Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.898926 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" event={"ID":"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f","Type":"ContainerDied","Data":"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.898965 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" event={"ID":"a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f","Type":"ContainerDied","Data":"50c6d198df9770621577a41f2964020a19e1f2605365ab7a9b9ed45ae8e3427b"} Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.898991 4861 scope.go:117] "RemoveContainer" containerID="fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.899188 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-7xg99" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.921132 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ccgz7" podStartSLOduration=2.499466221 podStartE2EDuration="3.921108861s" podCreationTimestamp="2026-03-15 00:15:11 +0000 UTC" firstStartedPulling="2026-03-15 00:15:12.84345851 +0000 UTC m=+506.155656043" lastFinishedPulling="2026-03-15 00:15:14.26510112 +0000 UTC m=+507.577298683" observedRunningTime="2026-03-15 00:15:14.916363677 +0000 UTC m=+508.228561230" watchObservedRunningTime="2026-03-15 00:15:14.921108861 +0000 UTC m=+508.233306394" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.931960 4861 scope.go:117] "RemoveContainer" containerID="fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea" Mar 15 00:15:14 crc kubenswrapper[4861]: E0315 00:15:14.933772 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea\": container with ID starting with fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea not found: ID does not exist" containerID="fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.933817 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea"} err="failed to get container status \"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea\": rpc error: code = NotFound desc = could not find container \"fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea\": container with ID starting with fe3c040f1221d1e66df6cf50bd7137192dfe6085e4c52befe018f1da30ef90ea not found: ID does not exist" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.939751 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.955721 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-7xg99"] Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959718 4861 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959757 4861 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959773 4861 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959785 4861 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959799 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhr8f\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-kube-api-access-xhr8f\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959810 4861 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:14 crc kubenswrapper[4861]: I0315 00:15:14.959822 4861 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.418346 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" path="/var/lib/kubelet/pods/a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f/volumes" Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.908097 4861 generic.go:334] "Generic (PLEG): container finished" podID="47cc9481-9f7a-47f2-933a-93bb592e1bdf" containerID="fbbc790822f3a78623e87e50c0aa84d70e8d39f074e5262aaea4ccc8355281ff" exitCode=0 Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.908230 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzwlb" event={"ID":"47cc9481-9f7a-47f2-933a-93bb592e1bdf","Type":"ContainerDied","Data":"fbbc790822f3a78623e87e50c0aa84d70e8d39f074e5262aaea4ccc8355281ff"} Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.913659 4861 generic.go:334] "Generic (PLEG): container finished" podID="f5a24ae0-68ec-4e10-9592-fa33d3380453" containerID="8b1cde3ef5950f4bcb3f3267dc8b491e45091d869b4a4f04da966ac587b896fa" exitCode=0 Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.913750 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzm9v" event={"ID":"f5a24ae0-68ec-4e10-9592-fa33d3380453","Type":"ContainerDied","Data":"8b1cde3ef5950f4bcb3f3267dc8b491e45091d869b4a4f04da966ac587b896fa"} Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.917084 4861 generic.go:334] "Generic (PLEG): container finished" podID="09972fc0-e7fc-4cd4-87d0-50d858ed9e51" containerID="593801c14b7765997a007c93f5d869fb8c1a255b300ea451439e6ca2cbbd311e" exitCode=0 Mar 15 00:15:15 crc kubenswrapper[4861]: I0315 00:15:15.917184 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w995g" event={"ID":"09972fc0-e7fc-4cd4-87d0-50d858ed9e51","Type":"ContainerDied","Data":"593801c14b7765997a007c93f5d869fb8c1a255b300ea451439e6ca2cbbd311e"} Mar 15 00:15:16 crc kubenswrapper[4861]: I0315 00:15:16.931294 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzm9v" event={"ID":"f5a24ae0-68ec-4e10-9592-fa33d3380453","Type":"ContainerStarted","Data":"d509f0b17f9dde3299fa6646fce232df1988a582b3879d348df2b5ad8b0c0012"} Mar 15 00:15:16 crc kubenswrapper[4861]: I0315 00:15:16.936164 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w995g" event={"ID":"09972fc0-e7fc-4cd4-87d0-50d858ed9e51","Type":"ContainerStarted","Data":"905c8c012d8381cd945d8da11766cc6937be83c7137bd1ddc98e27097e038b26"} Mar 15 00:15:16 crc kubenswrapper[4861]: I0315 00:15:16.944826 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzwlb" event={"ID":"47cc9481-9f7a-47f2-933a-93bb592e1bdf","Type":"ContainerStarted","Data":"7ca83b0dad446bb344ea01efe499f978e433160a5b37f264850ea81728a9d6c2"} Mar 15 00:15:16 crc kubenswrapper[4861]: I0315 00:15:16.993419 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pzwlb" podStartSLOduration=2.513874173 podStartE2EDuration="3.993383611s" podCreationTimestamp="2026-03-15 00:15:13 +0000 UTC" firstStartedPulling="2026-03-15 00:15:14.882132922 +0000 UTC m=+508.194330455" lastFinishedPulling="2026-03-15 00:15:16.36164236 +0000 UTC m=+509.673839893" observedRunningTime="2026-03-15 00:15:16.984885787 +0000 UTC m=+510.297083360" watchObservedRunningTime="2026-03-15 00:15:16.993383611 +0000 UTC m=+510.305581224" Mar 15 00:15:17 crc kubenswrapper[4861]: I0315 00:15:17.026802 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w995g" podStartSLOduration=3.549832188 podStartE2EDuration="6.026776463s" podCreationTimestamp="2026-03-15 00:15:11 +0000 UTC" firstStartedPulling="2026-03-15 00:15:13.867782998 +0000 UTC m=+507.179980531" lastFinishedPulling="2026-03-15 00:15:16.344727253 +0000 UTC m=+509.656924806" observedRunningTime="2026-03-15 00:15:17.01453869 +0000 UTC m=+510.326736293" watchObservedRunningTime="2026-03-15 00:15:17.026776463 +0000 UTC m=+510.338973996" Mar 15 00:15:17 crc kubenswrapper[4861]: I0315 00:15:17.961044 4861 generic.go:334] "Generic (PLEG): container finished" podID="f5a24ae0-68ec-4e10-9592-fa33d3380453" containerID="d509f0b17f9dde3299fa6646fce232df1988a582b3879d348df2b5ad8b0c0012" exitCode=0 Mar 15 00:15:17 crc kubenswrapper[4861]: I0315 00:15:17.962120 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzm9v" event={"ID":"f5a24ae0-68ec-4e10-9592-fa33d3380453","Type":"ContainerDied","Data":"d509f0b17f9dde3299fa6646fce232df1988a582b3879d348df2b5ad8b0c0012"} Mar 15 00:15:18 crc kubenswrapper[4861]: I0315 00:15:18.971205 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzm9v" event={"ID":"f5a24ae0-68ec-4e10-9592-fa33d3380453","Type":"ContainerStarted","Data":"edc2ef71263e28dbe3b9f0486b3a49a558e5074ad7e2a37085cae8a424780de6"} Mar 15 00:15:21 crc kubenswrapper[4861]: I0315 00:15:21.724810 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:21 crc kubenswrapper[4861]: I0315 00:15:21.725333 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:21 crc kubenswrapper[4861]: I0315 00:15:21.808517 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:21 crc kubenswrapper[4861]: I0315 00:15:21.828649 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qzm9v" podStartSLOduration=6.384679985 podStartE2EDuration="8.828621679s" podCreationTimestamp="2026-03-15 00:15:13 +0000 UTC" firstStartedPulling="2026-03-15 00:15:15.915068988 +0000 UTC m=+509.227266551" lastFinishedPulling="2026-03-15 00:15:18.359010712 +0000 UTC m=+511.671208245" observedRunningTime="2026-03-15 00:15:18.99060876 +0000 UTC m=+512.302806313" watchObservedRunningTime="2026-03-15 00:15:21.828621679 +0000 UTC m=+515.140819212" Mar 15 00:15:22 crc kubenswrapper[4861]: I0315 00:15:22.065098 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:15:22 crc kubenswrapper[4861]: I0315 00:15:22.907055 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:22 crc kubenswrapper[4861]: I0315 00:15:22.907159 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:23 crc kubenswrapper[4861]: I0315 00:15:23.978138 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w995g" podUID="09972fc0-e7fc-4cd4-87d0-50d858ed9e51" containerName="registry-server" probeResult="failure" output=< Mar 15 00:15:23 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:15:23 crc kubenswrapper[4861]: > Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.184413 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.184506 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.256939 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.318808 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.318920 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:24 crc kubenswrapper[4861]: I0315 00:15:24.387291 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:25 crc kubenswrapper[4861]: I0315 00:15:25.080515 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qzm9v" Mar 15 00:15:25 crc kubenswrapper[4861]: I0315 00:15:25.085140 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pzwlb" Mar 15 00:15:32 crc kubenswrapper[4861]: I0315 00:15:32.960173 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:15:33 crc kubenswrapper[4861]: I0315 00:15:33.029978 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w995g" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.165958 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558896-8qxmj"] Mar 15 00:16:00 crc kubenswrapper[4861]: E0315 00:16:00.167166 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" containerName="registry" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.167196 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" containerName="registry" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.167491 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="a68df7b1-2a4a-472d-b4ab-eed85f3c6e1f" containerName="registry" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.168299 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.172189 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.172425 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.174106 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.175068 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558896-8qxmj"] Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.300789 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gxxz\" (UniqueName: \"kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz\") pod \"auto-csr-approver-29558896-8qxmj\" (UID: \"7f75e814-555e-4ac9-ac2e-db0d34c970fc\") " pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.401807 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gxxz\" (UniqueName: \"kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz\") pod \"auto-csr-approver-29558896-8qxmj\" (UID: \"7f75e814-555e-4ac9-ac2e-db0d34c970fc\") " pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.431808 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gxxz\" (UniqueName: \"kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz\") pod \"auto-csr-approver-29558896-8qxmj\" (UID: \"7f75e814-555e-4ac9-ac2e-db0d34c970fc\") " pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.497041 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.749685 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558896-8qxmj"] Mar 15 00:16:00 crc kubenswrapper[4861]: I0315 00:16:00.761332 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:16:01 crc kubenswrapper[4861]: I0315 00:16:01.273044 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" event={"ID":"7f75e814-555e-4ac9-ac2e-db0d34c970fc","Type":"ContainerStarted","Data":"9960a1a6f7f308c189c7996ed2a3fcd7516e60fd5da2402732ee41c790207935"} Mar 15 00:16:02 crc kubenswrapper[4861]: I0315 00:16:02.281932 4861 generic.go:334] "Generic (PLEG): container finished" podID="7f75e814-555e-4ac9-ac2e-db0d34c970fc" containerID="de14f8b2aee9969d453d844e0e5179d2b0bbb786caf1f0d2af429ebb9853091e" exitCode=0 Mar 15 00:16:02 crc kubenswrapper[4861]: I0315 00:16:02.282010 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" event={"ID":"7f75e814-555e-4ac9-ac2e-db0d34c970fc","Type":"ContainerDied","Data":"de14f8b2aee9969d453d844e0e5179d2b0bbb786caf1f0d2af429ebb9853091e"} Mar 15 00:16:03 crc kubenswrapper[4861]: I0315 00:16:03.611875 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:03 crc kubenswrapper[4861]: I0315 00:16:03.754895 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gxxz\" (UniqueName: \"kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz\") pod \"7f75e814-555e-4ac9-ac2e-db0d34c970fc\" (UID: \"7f75e814-555e-4ac9-ac2e-db0d34c970fc\") " Mar 15 00:16:03 crc kubenswrapper[4861]: I0315 00:16:03.767148 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz" (OuterVolumeSpecName: "kube-api-access-9gxxz") pod "7f75e814-555e-4ac9-ac2e-db0d34c970fc" (UID: "7f75e814-555e-4ac9-ac2e-db0d34c970fc"). InnerVolumeSpecName "kube-api-access-9gxxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:16:03 crc kubenswrapper[4861]: I0315 00:16:03.856960 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gxxz\" (UniqueName: \"kubernetes.io/projected/7f75e814-555e-4ac9-ac2e-db0d34c970fc-kube-api-access-9gxxz\") on node \"crc\" DevicePath \"\"" Mar 15 00:16:04 crc kubenswrapper[4861]: I0315 00:16:04.304825 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" event={"ID":"7f75e814-555e-4ac9-ac2e-db0d34c970fc","Type":"ContainerDied","Data":"9960a1a6f7f308c189c7996ed2a3fcd7516e60fd5da2402732ee41c790207935"} Mar 15 00:16:04 crc kubenswrapper[4861]: I0315 00:16:04.304897 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9960a1a6f7f308c189c7996ed2a3fcd7516e60fd5da2402732ee41c790207935" Mar 15 00:16:04 crc kubenswrapper[4861]: I0315 00:16:04.305026 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558896-8qxmj" Mar 15 00:16:04 crc kubenswrapper[4861]: I0315 00:16:04.684924 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558890-7spp4"] Mar 15 00:16:04 crc kubenswrapper[4861]: I0315 00:16:04.689707 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558890-7spp4"] Mar 15 00:16:05 crc kubenswrapper[4861]: I0315 00:16:05.421229 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4585669d-fc84-4c80-9cc3-0185b7fb9ce0" path="/var/lib/kubelet/pods/4585669d-fc84-4c80-9cc3-0185b7fb9ce0/volumes" Mar 15 00:17:02 crc kubenswrapper[4861]: I0315 00:17:02.282664 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:17:02 crc kubenswrapper[4861]: I0315 00:17:02.283707 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:17:32 crc kubenswrapper[4861]: I0315 00:17:32.281904 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:17:32 crc kubenswrapper[4861]: I0315 00:17:32.282690 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.160507 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558898-ptwpx"] Mar 15 00:18:00 crc kubenswrapper[4861]: E0315 00:18:00.161833 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f75e814-555e-4ac9-ac2e-db0d34c970fc" containerName="oc" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.161855 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f75e814-555e-4ac9-ac2e-db0d34c970fc" containerName="oc" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.162021 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f75e814-555e-4ac9-ac2e-db0d34c970fc" containerName="oc" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.162655 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.169654 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.170072 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.171464 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.182424 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558898-ptwpx"] Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.249657 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kc8m\" (UniqueName: \"kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m\") pod \"auto-csr-approver-29558898-ptwpx\" (UID: \"6b7ede26-54ec-44c7-ad4d-81e7bccb0251\") " pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.351179 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kc8m\" (UniqueName: \"kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m\") pod \"auto-csr-approver-29558898-ptwpx\" (UID: \"6b7ede26-54ec-44c7-ad4d-81e7bccb0251\") " pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.381927 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kc8m\" (UniqueName: \"kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m\") pod \"auto-csr-approver-29558898-ptwpx\" (UID: \"6b7ede26-54ec-44c7-ad4d-81e7bccb0251\") " pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.487630 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:00 crc kubenswrapper[4861]: I0315 00:18:00.947223 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558898-ptwpx"] Mar 15 00:18:01 crc kubenswrapper[4861]: I0315 00:18:01.308074 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" event={"ID":"6b7ede26-54ec-44c7-ad4d-81e7bccb0251","Type":"ContainerStarted","Data":"26c25076f020a9072d6d77553b53dc12b608b077f90241d1857414f635ebfe4b"} Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.282699 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.283612 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.283849 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.284925 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.285234 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e" gracePeriod=600 Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.317275 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" event={"ID":"6b7ede26-54ec-44c7-ad4d-81e7bccb0251","Type":"ContainerStarted","Data":"5456683434de2f72fd984637411a566af681cd1a7101a6b34dcb99112b350e91"} Mar 15 00:18:02 crc kubenswrapper[4861]: I0315 00:18:02.345411 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" podStartSLOduration=1.496747034 podStartE2EDuration="2.345383997s" podCreationTimestamp="2026-03-15 00:18:00 +0000 UTC" firstStartedPulling="2026-03-15 00:18:00.960588831 +0000 UTC m=+674.272786404" lastFinishedPulling="2026-03-15 00:18:01.809225834 +0000 UTC m=+675.121423367" observedRunningTime="2026-03-15 00:18:02.338903924 +0000 UTC m=+675.651101477" watchObservedRunningTime="2026-03-15 00:18:02.345383997 +0000 UTC m=+675.657581550" Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.329470 4861 generic.go:334] "Generic (PLEG): container finished" podID="6b7ede26-54ec-44c7-ad4d-81e7bccb0251" containerID="5456683434de2f72fd984637411a566af681cd1a7101a6b34dcb99112b350e91" exitCode=0 Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.329658 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" event={"ID":"6b7ede26-54ec-44c7-ad4d-81e7bccb0251","Type":"ContainerDied","Data":"5456683434de2f72fd984637411a566af681cd1a7101a6b34dcb99112b350e91"} Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.335289 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e" exitCode=0 Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.335358 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e"} Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.335396 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0"} Mar 15 00:18:03 crc kubenswrapper[4861]: I0315 00:18:03.335418 4861 scope.go:117] "RemoveContainer" containerID="4e69150a8255080b9bcf4fc96f14b35a34722731237b6d673db953dbb97be3db" Mar 15 00:18:04 crc kubenswrapper[4861]: I0315 00:18:04.605401 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:04 crc kubenswrapper[4861]: I0315 00:18:04.645795 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kc8m\" (UniqueName: \"kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m\") pod \"6b7ede26-54ec-44c7-ad4d-81e7bccb0251\" (UID: \"6b7ede26-54ec-44c7-ad4d-81e7bccb0251\") " Mar 15 00:18:04 crc kubenswrapper[4861]: I0315 00:18:04.654573 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m" (OuterVolumeSpecName: "kube-api-access-2kc8m") pod "6b7ede26-54ec-44c7-ad4d-81e7bccb0251" (UID: "6b7ede26-54ec-44c7-ad4d-81e7bccb0251"). InnerVolumeSpecName "kube-api-access-2kc8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:18:04 crc kubenswrapper[4861]: I0315 00:18:04.746841 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kc8m\" (UniqueName: \"kubernetes.io/projected/6b7ede26-54ec-44c7-ad4d-81e7bccb0251-kube-api-access-2kc8m\") on node \"crc\" DevicePath \"\"" Mar 15 00:18:05 crc kubenswrapper[4861]: I0315 00:18:05.358444 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" event={"ID":"6b7ede26-54ec-44c7-ad4d-81e7bccb0251","Type":"ContainerDied","Data":"26c25076f020a9072d6d77553b53dc12b608b077f90241d1857414f635ebfe4b"} Mar 15 00:18:05 crc kubenswrapper[4861]: I0315 00:18:05.358499 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26c25076f020a9072d6d77553b53dc12b608b077f90241d1857414f635ebfe4b" Mar 15 00:18:05 crc kubenswrapper[4861]: I0315 00:18:05.358509 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558898-ptwpx" Mar 15 00:18:05 crc kubenswrapper[4861]: I0315 00:18:05.417091 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558892-8vbws"] Mar 15 00:18:05 crc kubenswrapper[4861]: I0315 00:18:05.419106 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558892-8vbws"] Mar 15 00:18:07 crc kubenswrapper[4861]: I0315 00:18:07.438791 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9efc1aaf-9347-4ed0-96ea-efa5342aa0b5" path="/var/lib/kubelet/pods/9efc1aaf-9347-4ed0-96ea-efa5342aa0b5/volumes" Mar 15 00:18:47 crc kubenswrapper[4861]: I0315 00:18:47.739148 4861 scope.go:117] "RemoveContainer" containerID="b303a70c6a8191095300a76be411edd13d0c8cbf766fd935a1db3236689f6d3e" Mar 15 00:18:47 crc kubenswrapper[4861]: I0315 00:18:47.804622 4861 scope.go:117] "RemoveContainer" containerID="43783405a59596ff4b17a60a37253b280aa04a127834d2ce8745dcc3108822c9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.406857 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pmf25"] Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.409250 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-controller" containerID="cri-o://818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.409995 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="northd" containerID="cri-o://517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.410520 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="sbdb" containerID="cri-o://15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.410713 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-acl-logging" containerID="cri-o://11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.410738 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="nbdb" containerID="cri-o://7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.410884 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.410969 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-node" containerID="cri-o://67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.478375 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" containerID="cri-o://2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" gracePeriod=30 Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.783863 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/3.log" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.787062 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovn-acl-logging/0.log" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.787922 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovn-controller/0.log" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.788673 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845447 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4f7n9"] Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845742 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-node" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845760 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-node" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845783 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845794 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845805 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="northd" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845814 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="northd" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845828 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="sbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845837 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="sbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845848 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7ede26-54ec-44c7-ad4d-81e7bccb0251" containerName="oc" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845858 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7ede26-54ec-44c7-ad4d-81e7bccb0251" containerName="oc" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845872 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845880 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845892 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="nbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845901 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="nbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845912 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845921 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845931 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845941 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845951 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-acl-logging" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845961 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-acl-logging" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.845975 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kubecfg-setup" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.845983 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kubecfg-setup" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.846000 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-ovn-metrics" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846009 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-ovn-metrics" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846134 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="sbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846236 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846247 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="nbdb" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846259 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846268 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846302 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-acl-logging" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846310 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846321 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-node" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846331 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7ede26-54ec-44c7-ad4d-81e7bccb0251" containerName="oc" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846341 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="northd" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846351 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovn-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846361 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="kube-rbac-proxy-ovn-metrics" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846373 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.846494 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846506 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: E0315 00:19:45.846770 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.846781 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerName="ovnkube-controller" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.848811 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862017 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862083 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862123 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862158 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862187 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862214 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862240 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862261 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862283 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862308 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k452t\" (UniqueName: \"kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862334 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862357 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862383 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862404 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862422 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862445 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862480 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862493 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket" (OuterVolumeSpecName: "log-socket") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862510 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862586 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862625 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862634 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862671 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch\") pod \"fa2d91cf-7d1a-4567-9593-654417784ef1\" (UID: \"fa2d91cf-7d1a-4567-9593-654417784ef1\") " Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862778 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862853 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862863 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-systemd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862890 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log" (OuterVolumeSpecName: "node-log") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862914 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-netns\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862921 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862945 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-var-lib-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.862976 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-log-socket\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863008 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863039 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48cae961-fd10-458e-bc59-26d7be73921b-ovn-node-metrics-cert\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863093 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863125 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-bin\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863209 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-kubelet\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863251 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-systemd-units\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863300 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-script-lib\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863346 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863417 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863433 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863425 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-ovn\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863465 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863484 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863503 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash" (OuterVolumeSpecName: "host-slash") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863511 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-node-log\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863523 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863547 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863670 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-etc-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863754 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8mqb\" (UniqueName: \"kubernetes.io/projected/48cae961-fd10-458e-bc59-26d7be73921b-kube-api-access-w8mqb\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863776 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-env-overrides\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863790 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863857 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-config\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863890 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-netd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863960 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.863996 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-slash\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864107 4861 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864131 4861 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-slash\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864149 4861 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864165 4861 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864177 4861 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-log-socket\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864190 4861 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864201 4861 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-node-log\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864215 4861 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864228 4861 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864243 4861 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864255 4861 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864269 4861 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864283 4861 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864296 4861 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.864308 4861 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.865522 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.873550 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t" (OuterVolumeSpecName: "kube-api-access-k452t") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "kube-api-access-k452t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.876600 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.888217 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "fa2d91cf-7d1a-4567-9593-654417784ef1" (UID: "fa2d91cf-7d1a-4567-9593-654417784ef1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.965859 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-netns\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966208 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-var-lib-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966302 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-var-lib-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966049 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-netns\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966321 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-log-socket\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966478 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-log-socket\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966617 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966721 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48cae961-fd10-458e-bc59-26d7be73921b-ovn-node-metrics-cert\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966623 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966881 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-bin\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966970 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967074 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-kubelet\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967141 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967125 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-kubelet\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967157 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-systemd-units\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967311 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-script-lib\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967387 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967462 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-ovn\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.966909 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-bin\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967492 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-run-ovn-kubernetes\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967542 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-node-log\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967545 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-ovn\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967506 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-node-log\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967742 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-etc-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967824 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-systemd-units\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967828 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8mqb\" (UniqueName: \"kubernetes.io/projected/48cae961-fd10-458e-bc59-26d7be73921b-kube-api-access-w8mqb\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968008 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-env-overrides\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.967921 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-etc-openvswitch\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968158 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-netd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968223 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-cni-netd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968307 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-config\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968405 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-slash\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968480 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-systemd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968579 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-host-slash\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968632 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/48cae961-fd10-458e-bc59-26d7be73921b-run-systemd\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968601 4861 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fa2d91cf-7d1a-4567-9593-654417784ef1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968724 4861 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968748 4861 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fa2d91cf-7d1a-4567-9593-654417784ef1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968767 4861 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fa2d91cf-7d1a-4567-9593-654417784ef1-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.968782 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k452t\" (UniqueName: \"kubernetes.io/projected/fa2d91cf-7d1a-4567-9593-654417784ef1-kube-api-access-k452t\") on node \"crc\" DevicePath \"\"" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.969101 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-env-overrides\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.969135 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-script-lib\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.969216 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/48cae961-fd10-458e-bc59-26d7be73921b-ovnkube-config\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.970369 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/48cae961-fd10-458e-bc59-26d7be73921b-ovn-node-metrics-cert\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:45 crc kubenswrapper[4861]: I0315 00:19:45.988137 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8mqb\" (UniqueName: \"kubernetes.io/projected/48cae961-fd10-458e-bc59-26d7be73921b-kube-api-access-w8mqb\") pod \"ovnkube-node-4f7n9\" (UID: \"48cae961-fd10-458e-bc59-26d7be73921b\") " pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.094054 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovnkube-controller/3.log" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.098084 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovn-acl-logging/0.log" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.098815 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pmf25_fa2d91cf-7d1a-4567-9593-654417784ef1/ovn-controller/0.log" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099296 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099352 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099370 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099384 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099397 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099410 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" exitCode=0 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099421 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" exitCode=143 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099422 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099465 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099434 4861 generic.go:334] "Generic (PLEG): container finished" podID="fa2d91cf-7d1a-4567-9593-654417784ef1" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" exitCode=143 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099609 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099653 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099683 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099687 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099717 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099747 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099779 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099799 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099813 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099826 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099839 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099852 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099866 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099880 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099895 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099912 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099935 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099954 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099970 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.099985 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100001 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100017 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100034 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100051 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100068 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100085 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100106 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100130 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100148 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100162 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100177 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100192 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100208 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100222 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100238 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100253 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100268 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100291 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pmf25" event={"ID":"fa2d91cf-7d1a-4567-9593-654417784ef1","Type":"ContainerDied","Data":"fed8b2ae383c8884aa99f085a158f86ff2a696c61c4d1e82d0ed69f359da793d"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100320 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100339 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100355 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100371 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100390 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100408 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100425 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100440 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100456 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.100472 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.104614 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/2.log" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.105759 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/1.log" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.105844 4861 generic.go:334] "Generic (PLEG): container finished" podID="8cdc187d-d699-431a-9355-5a2268bf3821" containerID="4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746" exitCode=2 Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.105901 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerDied","Data":"4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.105939 4861 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f"} Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.107016 4861 scope.go:117] "RemoveContainer" containerID="4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.107464 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821)\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.139360 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.164972 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.178964 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pmf25"] Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.191597 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pmf25"] Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.192889 4861 scope.go:117] "RemoveContainer" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.215966 4861 scope.go:117] "RemoveContainer" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.236951 4861 scope.go:117] "RemoveContainer" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.256448 4861 scope.go:117] "RemoveContainer" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.280521 4861 scope.go:117] "RemoveContainer" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.305165 4861 scope.go:117] "RemoveContainer" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.333546 4861 scope.go:117] "RemoveContainer" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.358392 4861 scope.go:117] "RemoveContainer" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.406290 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.407917 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.407953 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} err="failed to get container status \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.407979 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.409674 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": container with ID starting with 87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0 not found: ID does not exist" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.409752 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} err="failed to get container status \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": rpc error: code = NotFound desc = could not find container \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": container with ID starting with 87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.409769 4861 scope.go:117] "RemoveContainer" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.410079 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": container with ID starting with 15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264 not found: ID does not exist" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410097 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} err="failed to get container status \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": rpc error: code = NotFound desc = could not find container \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": container with ID starting with 15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410109 4861 scope.go:117] "RemoveContainer" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.410413 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": container with ID starting with 7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405 not found: ID does not exist" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410430 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} err="failed to get container status \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": rpc error: code = NotFound desc = could not find container \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": container with ID starting with 7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410442 4861 scope.go:117] "RemoveContainer" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.410693 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": container with ID starting with 517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77 not found: ID does not exist" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410732 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} err="failed to get container status \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": rpc error: code = NotFound desc = could not find container \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": container with ID starting with 517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410744 4861 scope.go:117] "RemoveContainer" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.410936 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": container with ID starting with 72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f not found: ID does not exist" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.410968 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} err="failed to get container status \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": rpc error: code = NotFound desc = could not find container \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": container with ID starting with 72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411022 4861 scope.go:117] "RemoveContainer" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.411221 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": container with ID starting with 67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141 not found: ID does not exist" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411237 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} err="failed to get container status \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": rpc error: code = NotFound desc = could not find container \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": container with ID starting with 67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411249 4861 scope.go:117] "RemoveContainer" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.411427 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": container with ID starting with 11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da not found: ID does not exist" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411440 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} err="failed to get container status \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": rpc error: code = NotFound desc = could not find container \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": container with ID starting with 11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411454 4861 scope.go:117] "RemoveContainer" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.411679 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": container with ID starting with 818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca not found: ID does not exist" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411703 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} err="failed to get container status \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": rpc error: code = NotFound desc = could not find container \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": container with ID starting with 818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411717 4861 scope.go:117] "RemoveContainer" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: E0315 00:19:46.411900 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": container with ID starting with a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416 not found: ID does not exist" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411914 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} err="failed to get container status \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": rpc error: code = NotFound desc = could not find container \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": container with ID starting with a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.411926 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412143 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} err="failed to get container status \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412187 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412360 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} err="failed to get container status \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": rpc error: code = NotFound desc = could not find container \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": container with ID starting with 87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412376 4861 scope.go:117] "RemoveContainer" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412531 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} err="failed to get container status \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": rpc error: code = NotFound desc = could not find container \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": container with ID starting with 15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412544 4861 scope.go:117] "RemoveContainer" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412743 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} err="failed to get container status \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": rpc error: code = NotFound desc = could not find container \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": container with ID starting with 7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412764 4861 scope.go:117] "RemoveContainer" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412964 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} err="failed to get container status \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": rpc error: code = NotFound desc = could not find container \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": container with ID starting with 517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.412982 4861 scope.go:117] "RemoveContainer" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.413438 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} err="failed to get container status \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": rpc error: code = NotFound desc = could not find container \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": container with ID starting with 72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.413457 4861 scope.go:117] "RemoveContainer" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414127 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} err="failed to get container status \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": rpc error: code = NotFound desc = could not find container \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": container with ID starting with 67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414147 4861 scope.go:117] "RemoveContainer" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414466 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} err="failed to get container status \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": rpc error: code = NotFound desc = could not find container \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": container with ID starting with 11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414486 4861 scope.go:117] "RemoveContainer" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414844 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} err="failed to get container status \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": rpc error: code = NotFound desc = could not find container \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": container with ID starting with 818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.414871 4861 scope.go:117] "RemoveContainer" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.415195 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} err="failed to get container status \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": rpc error: code = NotFound desc = could not find container \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": container with ID starting with a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.415215 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.415681 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} err="failed to get container status \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.415700 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.416175 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} err="failed to get container status \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": rpc error: code = NotFound desc = could not find container \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": container with ID starting with 87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.416196 4861 scope.go:117] "RemoveContainer" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.416742 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} err="failed to get container status \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": rpc error: code = NotFound desc = could not find container \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": container with ID starting with 15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.416764 4861 scope.go:117] "RemoveContainer" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.417025 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} err="failed to get container status \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": rpc error: code = NotFound desc = could not find container \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": container with ID starting with 7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.417049 4861 scope.go:117] "RemoveContainer" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.417463 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} err="failed to get container status \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": rpc error: code = NotFound desc = could not find container \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": container with ID starting with 517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.417484 4861 scope.go:117] "RemoveContainer" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.418101 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} err="failed to get container status \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": rpc error: code = NotFound desc = could not find container \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": container with ID starting with 72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.418120 4861 scope.go:117] "RemoveContainer" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.418546 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} err="failed to get container status \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": rpc error: code = NotFound desc = could not find container \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": container with ID starting with 67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.418592 4861 scope.go:117] "RemoveContainer" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.419088 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} err="failed to get container status \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": rpc error: code = NotFound desc = could not find container \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": container with ID starting with 11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.419155 4861 scope.go:117] "RemoveContainer" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.419569 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} err="failed to get container status \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": rpc error: code = NotFound desc = could not find container \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": container with ID starting with 818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.419597 4861 scope.go:117] "RemoveContainer" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420073 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} err="failed to get container status \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": rpc error: code = NotFound desc = could not find container \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": container with ID starting with a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420102 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420374 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} err="failed to get container status \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420412 4861 scope.go:117] "RemoveContainer" containerID="87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420821 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0"} err="failed to get container status \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": rpc error: code = NotFound desc = could not find container \"87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0\": container with ID starting with 87bd8067c60738168e00a811294f32a977cd2dc49452b09e4261404dca6dabc0 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.420843 4861 scope.go:117] "RemoveContainer" containerID="15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.421153 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264"} err="failed to get container status \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": rpc error: code = NotFound desc = could not find container \"15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264\": container with ID starting with 15a7c4d12332c0eaaf65816703e1d6b16961d2e66a8fadb10c3dbe2da5ab4264 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.421191 4861 scope.go:117] "RemoveContainer" containerID="7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.421838 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405"} err="failed to get container status \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": rpc error: code = NotFound desc = could not find container \"7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405\": container with ID starting with 7443923a4f1dc957a24ab803cf60ee0c1a8007d23e3ea4c93bfe2372b8e92405 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.421859 4861 scope.go:117] "RemoveContainer" containerID="517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.422317 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77"} err="failed to get container status \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": rpc error: code = NotFound desc = could not find container \"517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77\": container with ID starting with 517fb2746bd955a7aa17e0a12511663db6fead56ae87fcd3c383dba867d6fe77 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.422344 4861 scope.go:117] "RemoveContainer" containerID="72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.422665 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f"} err="failed to get container status \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": rpc error: code = NotFound desc = could not find container \"72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f\": container with ID starting with 72e5f749bc0d65a8926167bd9594b72f7deb2ca2c854123f788eb76d74c0e15f not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.422703 4861 scope.go:117] "RemoveContainer" containerID="67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423009 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141"} err="failed to get container status \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": rpc error: code = NotFound desc = could not find container \"67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141\": container with ID starting with 67e1601cd5a8e7f7f93541a48f78f78a6244ae926b36e4808cb5ad68a17b1141 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423041 4861 scope.go:117] "RemoveContainer" containerID="11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423269 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da"} err="failed to get container status \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": rpc error: code = NotFound desc = could not find container \"11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da\": container with ID starting with 11dcab5cee556861ad420a22ddc431951b042745ad72aea68e89dac4c4acb4da not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423286 4861 scope.go:117] "RemoveContainer" containerID="818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423518 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca"} err="failed to get container status \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": rpc error: code = NotFound desc = could not find container \"818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca\": container with ID starting with 818327f0b3bebd938e734c6a532e6402ae8984e68ecb098ac41f5cdac78c2bca not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423580 4861 scope.go:117] "RemoveContainer" containerID="a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423848 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416"} err="failed to get container status \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": rpc error: code = NotFound desc = could not find container \"a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416\": container with ID starting with a95a0112fcd44ca56c6c375f2bc58306d801ea0b5551ec4760a01c9f3466f416 not found: ID does not exist" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.423878 4861 scope.go:117] "RemoveContainer" containerID="2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b" Mar 15 00:19:46 crc kubenswrapper[4861]: I0315 00:19:46.424109 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b"} err="failed to get container status \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": rpc error: code = NotFound desc = could not find container \"2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b\": container with ID starting with 2361489f17646335b55e59e6099b6d8df99528bcfea215cfa8723838a1d0800b not found: ID does not exist" Mar 15 00:19:47 crc kubenswrapper[4861]: I0315 00:19:47.117454 4861 generic.go:334] "Generic (PLEG): container finished" podID="48cae961-fd10-458e-bc59-26d7be73921b" containerID="176d1e9de9c3a1e79e4110b2467a1e2a273c580af24cabee79eb05f5161629e7" exitCode=0 Mar 15 00:19:47 crc kubenswrapper[4861]: I0315 00:19:47.117596 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerDied","Data":"176d1e9de9c3a1e79e4110b2467a1e2a273c580af24cabee79eb05f5161629e7"} Mar 15 00:19:47 crc kubenswrapper[4861]: I0315 00:19:47.117641 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"e51f2d2d0eec4fc9cb6513563551222e334df7f8db52d983d2f142692eaa30c6"} Mar 15 00:19:47 crc kubenswrapper[4861]: I0315 00:19:47.431328 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa2d91cf-7d1a-4567-9593-654417784ef1" path="/var/lib/kubelet/pods/fa2d91cf-7d1a-4567-9593-654417784ef1/volumes" Mar 15 00:19:47 crc kubenswrapper[4861]: I0315 00:19:47.901354 4861 scope.go:117] "RemoveContainer" containerID="671289f6abb1a7b353009c25bc4ff09f1a302bb13d83cd7d331cb470fd2f5a6f" Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134303 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"5feb01035c082f7f25597f0633e6bdfddaad415e464dc289d14ab5baf1795e13"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134368 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"be47049d80e02969004d59a78886c7561f1dd2dabd3109f0aee2a46b3e406895"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134384 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"c50c6f569458474d813903f0218cf019ba6037b35f99671f2540b756a675c7c4"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134409 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"4485b602243ab32779c75e2b00812811e7aa9831716c3d1d095f2eefe2dc2d52"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134425 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"1659965d73ba55f60bd4d88330218d03ccf191dbd5c72b686b313a5ad172e011"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.134438 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"016c4ccca48007808a4a1a5b115539e44135e43ba3693bde57d7ca1b5146b7bc"} Mar 15 00:19:48 crc kubenswrapper[4861]: I0315 00:19:48.136165 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/2.log" Mar 15 00:19:51 crc kubenswrapper[4861]: I0315 00:19:51.163344 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"cff8bc5f5ac607edb6f0863ef12faa7e848c7f3eefb142090e64a933a582ec02"} Mar 15 00:19:53 crc kubenswrapper[4861]: I0315 00:19:53.188755 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" event={"ID":"48cae961-fd10-458e-bc59-26d7be73921b","Type":"ContainerStarted","Data":"be089ff674a08ccdfb2f91a76d8b91d748e3ae5ec996c760544b1af046b45523"} Mar 15 00:19:53 crc kubenswrapper[4861]: I0315 00:19:53.189172 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:53 crc kubenswrapper[4861]: I0315 00:19:53.189186 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:53 crc kubenswrapper[4861]: I0315 00:19:53.216970 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:53 crc kubenswrapper[4861]: I0315 00:19:53.219753 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" podStartSLOduration=8.219739775 podStartE2EDuration="8.219739775s" podCreationTimestamp="2026-03-15 00:19:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:19:53.218596867 +0000 UTC m=+786.530794400" watchObservedRunningTime="2026-03-15 00:19:53.219739775 +0000 UTC m=+786.531937308" Mar 15 00:19:54 crc kubenswrapper[4861]: I0315 00:19:54.199616 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:54 crc kubenswrapper[4861]: I0315 00:19:54.248166 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:19:57 crc kubenswrapper[4861]: I0315 00:19:57.413929 4861 scope.go:117] "RemoveContainer" containerID="4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746" Mar 15 00:19:57 crc kubenswrapper[4861]: E0315 00:19:57.414601 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bbzq6_openshift-multus(8cdc187d-d699-431a-9355-5a2268bf3821)\"" pod="openshift-multus/multus-bbzq6" podUID="8cdc187d-d699-431a-9355-5a2268bf3821" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.150132 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558900-dm476"] Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.151428 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.153923 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.153950 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.154219 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.212509 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558900-dm476"] Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.282245 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvpb\" (UniqueName: \"kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb\") pod \"auto-csr-approver-29558900-dm476\" (UID: \"34e35417-c6f9-47c0-84c9-0ef2f6482447\") " pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.384187 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvpb\" (UniqueName: \"kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb\") pod \"auto-csr-approver-29558900-dm476\" (UID: \"34e35417-c6f9-47c0-84c9-0ef2f6482447\") " pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.411387 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvpb\" (UniqueName: \"kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb\") pod \"auto-csr-approver-29558900-dm476\" (UID: \"34e35417-c6f9-47c0-84c9-0ef2f6482447\") " pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: I0315 00:20:00.481410 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: E0315 00:20:00.534490 4861 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(cbb812a953bfe3d52a79c29cdca7d55a0abae6c8c18a8b03c57ef415fbbdc6e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:20:00 crc kubenswrapper[4861]: E0315 00:20:00.534683 4861 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(cbb812a953bfe3d52a79c29cdca7d55a0abae6c8c18a8b03c57ef415fbbdc6e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: E0315 00:20:00.534739 4861 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(cbb812a953bfe3d52a79c29cdca7d55a0abae6c8c18a8b03c57ef415fbbdc6e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:00 crc kubenswrapper[4861]: E0315 00:20:00.534891 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29558900-dm476_openshift-infra(34e35417-c6f9-47c0-84c9-0ef2f6482447)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29558900-dm476_openshift-infra(34e35417-c6f9-47c0-84c9-0ef2f6482447)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(cbb812a953bfe3d52a79c29cdca7d55a0abae6c8c18a8b03c57ef415fbbdc6e1): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29558900-dm476" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" Mar 15 00:20:01 crc kubenswrapper[4861]: I0315 00:20:01.252117 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:01 crc kubenswrapper[4861]: I0315 00:20:01.254036 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:01 crc kubenswrapper[4861]: E0315 00:20:01.302522 4861 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(c335c57920fd405452ca948bbd3fc88e794c2fe8010932dad6145e74afa4cd06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 15 00:20:01 crc kubenswrapper[4861]: E0315 00:20:01.302615 4861 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(c335c57920fd405452ca948bbd3fc88e794c2fe8010932dad6145e74afa4cd06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:01 crc kubenswrapper[4861]: E0315 00:20:01.302645 4861 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(c335c57920fd405452ca948bbd3fc88e794c2fe8010932dad6145e74afa4cd06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:01 crc kubenswrapper[4861]: E0315 00:20:01.302712 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29558900-dm476_openshift-infra(34e35417-c6f9-47c0-84c9-0ef2f6482447)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29558900-dm476_openshift-infra(34e35417-c6f9-47c0-84c9-0ef2f6482447)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29558900-dm476_openshift-infra_34e35417-c6f9-47c0-84c9-0ef2f6482447_0(c335c57920fd405452ca948bbd3fc88e794c2fe8010932dad6145e74afa4cd06): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29558900-dm476" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" Mar 15 00:20:02 crc kubenswrapper[4861]: I0315 00:20:02.282931 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:20:02 crc kubenswrapper[4861]: I0315 00:20:02.283005 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:20:09 crc kubenswrapper[4861]: I0315 00:20:09.410078 4861 scope.go:117] "RemoveContainer" containerID="4cdc566711233a81478c582b5defa482de0beef5a18a78ceb131696d864ea746" Mar 15 00:20:10 crc kubenswrapper[4861]: I0315 00:20:10.331532 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bbzq6_8cdc187d-d699-431a-9355-5a2268bf3821/kube-multus/2.log" Mar 15 00:20:10 crc kubenswrapper[4861]: I0315 00:20:10.332118 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bbzq6" event={"ID":"8cdc187d-d699-431a-9355-5a2268bf3821","Type":"ContainerStarted","Data":"6f944e8039b854f084738e8624de76ed4f5e9ab7a7ccaa37deecc28ce650eadd"} Mar 15 00:20:15 crc kubenswrapper[4861]: I0315 00:20:15.408387 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:15 crc kubenswrapper[4861]: I0315 00:20:15.409611 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:15 crc kubenswrapper[4861]: I0315 00:20:15.692354 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558900-dm476"] Mar 15 00:20:16 crc kubenswrapper[4861]: I0315 00:20:16.197249 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4f7n9" Mar 15 00:20:16 crc kubenswrapper[4861]: I0315 00:20:16.373632 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558900-dm476" event={"ID":"34e35417-c6f9-47c0-84c9-0ef2f6482447","Type":"ContainerStarted","Data":"5508d3ac8f0e46bdad2e3c39c12a5b0baa60e9cec1f3a4d71d261e319c01c475"} Mar 15 00:20:18 crc kubenswrapper[4861]: I0315 00:20:18.399194 4861 generic.go:334] "Generic (PLEG): container finished" podID="34e35417-c6f9-47c0-84c9-0ef2f6482447" containerID="00551e8d857c4aa507b818b487aaef5d382d34ed4b821ee965749493f7db6537" exitCode=0 Mar 15 00:20:18 crc kubenswrapper[4861]: I0315 00:20:18.399281 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558900-dm476" event={"ID":"34e35417-c6f9-47c0-84c9-0ef2f6482447","Type":"ContainerDied","Data":"00551e8d857c4aa507b818b487aaef5d382d34ed4b821ee965749493f7db6537"} Mar 15 00:20:19 crc kubenswrapper[4861]: I0315 00:20:19.689373 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:19 crc kubenswrapper[4861]: I0315 00:20:19.873820 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmvpb\" (UniqueName: \"kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb\") pod \"34e35417-c6f9-47c0-84c9-0ef2f6482447\" (UID: \"34e35417-c6f9-47c0-84c9-0ef2f6482447\") " Mar 15 00:20:19 crc kubenswrapper[4861]: I0315 00:20:19.882546 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb" (OuterVolumeSpecName: "kube-api-access-tmvpb") pod "34e35417-c6f9-47c0-84c9-0ef2f6482447" (UID: "34e35417-c6f9-47c0-84c9-0ef2f6482447"). InnerVolumeSpecName "kube-api-access-tmvpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:20:19 crc kubenswrapper[4861]: I0315 00:20:19.975737 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmvpb\" (UniqueName: \"kubernetes.io/projected/34e35417-c6f9-47c0-84c9-0ef2f6482447-kube-api-access-tmvpb\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:20 crc kubenswrapper[4861]: I0315 00:20:20.417229 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558900-dm476" event={"ID":"34e35417-c6f9-47c0-84c9-0ef2f6482447","Type":"ContainerDied","Data":"5508d3ac8f0e46bdad2e3c39c12a5b0baa60e9cec1f3a4d71d261e319c01c475"} Mar 15 00:20:20 crc kubenswrapper[4861]: I0315 00:20:20.417321 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5508d3ac8f0e46bdad2e3c39c12a5b0baa60e9cec1f3a4d71d261e319c01c475" Mar 15 00:20:20 crc kubenswrapper[4861]: I0315 00:20:20.417329 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558900-dm476" Mar 15 00:20:20 crc kubenswrapper[4861]: I0315 00:20:20.773665 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558894-d7mzm"] Mar 15 00:20:20 crc kubenswrapper[4861]: I0315 00:20:20.783239 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558894-d7mzm"] Mar 15 00:20:21 crc kubenswrapper[4861]: I0315 00:20:21.418475 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a" path="/var/lib/kubelet/pods/d7630b8c-89ef-47f6-a9ff-a8924a3d7d0a/volumes" Mar 15 00:20:32 crc kubenswrapper[4861]: I0315 00:20:32.282771 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:20:32 crc kubenswrapper[4861]: I0315 00:20:32.283549 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.312346 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.313663 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ccgz7" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="registry-server" containerID="cri-o://f0f3c7320fdb655e7c54f68b728f4f820e835f6801d1f93f7b69b01240ea5f31" gracePeriod=30 Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.576440 4861 generic.go:334] "Generic (PLEG): container finished" podID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerID="f0f3c7320fdb655e7c54f68b728f4f820e835f6801d1f93f7b69b01240ea5f31" exitCode=0 Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.576938 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerDied","Data":"f0f3c7320fdb655e7c54f68b728f4f820e835f6801d1f93f7b69b01240ea5f31"} Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.676224 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.842349 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content\") pod \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.842408 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt6ch\" (UniqueName: \"kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch\") pod \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.842434 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities\") pod \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\" (UID: \"3fb39d55-b4c1-4baf-9d0b-efce980b11c9\") " Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.843732 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities" (OuterVolumeSpecName: "utilities") pod "3fb39d55-b4c1-4baf-9d0b-efce980b11c9" (UID: "3fb39d55-b4c1-4baf-9d0b-efce980b11c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.844304 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.849046 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch" (OuterVolumeSpecName: "kube-api-access-gt6ch") pod "3fb39d55-b4c1-4baf-9d0b-efce980b11c9" (UID: "3fb39d55-b4c1-4baf-9d0b-efce980b11c9"). InnerVolumeSpecName "kube-api-access-gt6ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.873343 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fb39d55-b4c1-4baf-9d0b-efce980b11c9" (UID: "3fb39d55-b4c1-4baf-9d0b-efce980b11c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.946351 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:41 crc kubenswrapper[4861]: I0315 00:20:41.946415 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt6ch\" (UniqueName: \"kubernetes.io/projected/3fb39d55-b4c1-4baf-9d0b-efce980b11c9-kube-api-access-gt6ch\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.589796 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ccgz7" event={"ID":"3fb39d55-b4c1-4baf-9d0b-efce980b11c9","Type":"ContainerDied","Data":"3d3f309dce4b8a9883b1c41b8260be3ba51b1f0c08d03ae65cb54abe3a343370"} Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.590392 4861 scope.go:117] "RemoveContainer" containerID="f0f3c7320fdb655e7c54f68b728f4f820e835f6801d1f93f7b69b01240ea5f31" Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.589936 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ccgz7" Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.618233 4861 scope.go:117] "RemoveContainer" containerID="d10f3bbaf7843210c8b1dd84f94ce86c1ef463868de551df4cbbf3d378a80ec0" Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.636359 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.640439 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ccgz7"] Mar 15 00:20:42 crc kubenswrapper[4861]: I0315 00:20:42.659314 4861 scope.go:117] "RemoveContainer" containerID="b20434563ff42a55c696bc28ab446ea819347465647c44425fe1043564a8b56f" Mar 15 00:20:43 crc kubenswrapper[4861]: I0315 00:20:43.419419 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" path="/var/lib/kubelet/pods/3fb39d55-b4c1-4baf-9d0b-efce980b11c9/volumes" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.277863 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn"] Mar 15 00:20:45 crc kubenswrapper[4861]: E0315 00:20:45.278385 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="extract-content" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278398 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="extract-content" Mar 15 00:20:45 crc kubenswrapper[4861]: E0315 00:20:45.278411 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" containerName="oc" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278419 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" containerName="oc" Mar 15 00:20:45 crc kubenswrapper[4861]: E0315 00:20:45.278428 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="extract-utilities" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278434 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="extract-utilities" Mar 15 00:20:45 crc kubenswrapper[4861]: E0315 00:20:45.278446 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="registry-server" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278452 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="registry-server" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278585 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" containerName="oc" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.278594 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb39d55-b4c1-4baf-9d0b-efce980b11c9" containerName="registry-server" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.279325 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.288495 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.292957 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn"] Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.394687 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.395688 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.395761 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfsmx\" (UniqueName: \"kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.497033 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.497154 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfsmx\" (UniqueName: \"kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.497302 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.498022 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.498719 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.534669 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfsmx\" (UniqueName: \"kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.595315 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:45 crc kubenswrapper[4861]: I0315 00:20:45.820199 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn"] Mar 15 00:20:46 crc kubenswrapper[4861]: I0315 00:20:46.618616 4861 generic.go:334] "Generic (PLEG): container finished" podID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerID="f668b59c9873458149443969a6fce15a2e828e1b884e9ed8dddef5040bde7855" exitCode=0 Mar 15 00:20:46 crc kubenswrapper[4861]: I0315 00:20:46.618688 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" event={"ID":"4e23db65-ec89-427d-aaa4-21d9c00779fd","Type":"ContainerDied","Data":"f668b59c9873458149443969a6fce15a2e828e1b884e9ed8dddef5040bde7855"} Mar 15 00:20:46 crc kubenswrapper[4861]: I0315 00:20:46.618728 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" event={"ID":"4e23db65-ec89-427d-aaa4-21d9c00779fd","Type":"ContainerStarted","Data":"c073354f0b6afa8db93bfcd71d7a215edcce28bcbecedf67e5db7438388e8ce8"} Mar 15 00:20:47 crc kubenswrapper[4861]: I0315 00:20:47.967536 4861 scope.go:117] "RemoveContainer" containerID="430ae6b9ad7d2ba2b6fdd32abf6b778ffb46c52025c3f506fcb15dfbfbf5523e" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.684144 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b"] Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.685846 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.697231 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b"] Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.789143 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.789249 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.789299 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khn84\" (UniqueName: \"kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.890647 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.890722 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.890752 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khn84\" (UniqueName: \"kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.891385 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.891798 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:50 crc kubenswrapper[4861]: I0315 00:20:50.939430 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khn84\" (UniqueName: \"kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.006394 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.298420 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b"] Mar 15 00:20:51 crc kubenswrapper[4861]: W0315 00:20:51.307787 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17925b27_9408_4067_970e_de025397edf2.slice/crio-565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3 WatchSource:0}: Error finding container 565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3: Status 404 returned error can't find the container with id 565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3 Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.473361 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs"] Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.474854 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.497739 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh6q8\" (UniqueName: \"kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.497813 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.497956 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.498873 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs"] Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.599659 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.600296 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh6q8\" (UniqueName: \"kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.600610 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.600705 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.601111 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.634801 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh6q8\" (UniqueName: \"kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.661417 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerStarted","Data":"450361a77cc431322e1f92c200f5a5c6b92192472fd6dcdd7a7e5946529f0aa0"} Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.661492 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerStarted","Data":"565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3"} Mar 15 00:20:51 crc kubenswrapper[4861]: I0315 00:20:51.797151 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.051548 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs"] Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.669251 4861 generic.go:334] "Generic (PLEG): container finished" podID="17925b27-9408-4067-970e-de025397edf2" containerID="450361a77cc431322e1f92c200f5a5c6b92192472fd6dcdd7a7e5946529f0aa0" exitCode=0 Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.669334 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerDied","Data":"450361a77cc431322e1f92c200f5a5c6b92192472fd6dcdd7a7e5946529f0aa0"} Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.674486 4861 generic.go:334] "Generic (PLEG): container finished" podID="2c747e57-e783-45d0-a353-521096017d15" containerID="b8c613b07760867e52e5e3aa567fef8a8bb6b13c70cffbaea3a64c47c0fa808c" exitCode=0 Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.674592 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" event={"ID":"2c747e57-e783-45d0-a353-521096017d15","Type":"ContainerDied","Data":"b8c613b07760867e52e5e3aa567fef8a8bb6b13c70cffbaea3a64c47c0fa808c"} Mar 15 00:20:52 crc kubenswrapper[4861]: I0315 00:20:52.674638 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" event={"ID":"2c747e57-e783-45d0-a353-521096017d15","Type":"ContainerStarted","Data":"e8c2df49701d2fa7e2aced354300c5c3547b33569c232acd9dfd59fedac19c18"} Mar 15 00:20:53 crc kubenswrapper[4861]: I0315 00:20:53.685704 4861 generic.go:334] "Generic (PLEG): container finished" podID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerID="e4553d12a20ab0547790cb7d95877a0498eed0f0f8490a365a8fb0b0ea1117c0" exitCode=0 Mar 15 00:20:53 crc kubenswrapper[4861]: I0315 00:20:53.685765 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" event={"ID":"4e23db65-ec89-427d-aaa4-21d9c00779fd","Type":"ContainerDied","Data":"e4553d12a20ab0547790cb7d95877a0498eed0f0f8490a365a8fb0b0ea1117c0"} Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.695691 4861 generic.go:334] "Generic (PLEG): container finished" podID="2c747e57-e783-45d0-a353-521096017d15" containerID="89ce925be90c72493cf5566b135b796387d914511ccce5e8856de1fc189f4a37" exitCode=0 Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.695748 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" event={"ID":"2c747e57-e783-45d0-a353-521096017d15","Type":"ContainerDied","Data":"89ce925be90c72493cf5566b135b796387d914511ccce5e8856de1fc189f4a37"} Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.703121 4861 generic.go:334] "Generic (PLEG): container finished" podID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerID="92d957be5bbe878bc37032d832e0081f45140514ba68c1ae99d50b82e2df6a90" exitCode=0 Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.703217 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" event={"ID":"4e23db65-ec89-427d-aaa4-21d9c00779fd","Type":"ContainerDied","Data":"92d957be5bbe878bc37032d832e0081f45140514ba68c1ae99d50b82e2df6a90"} Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.706653 4861 generic.go:334] "Generic (PLEG): container finished" podID="17925b27-9408-4067-970e-de025397edf2" containerID="88f70bf0a28060aacac3e46240c98621639bf77c016f1ce8f3b36f23424d8756" exitCode=0 Mar 15 00:20:54 crc kubenswrapper[4861]: I0315 00:20:54.706696 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerDied","Data":"88f70bf0a28060aacac3e46240c98621639bf77c016f1ce8f3b36f23424d8756"} Mar 15 00:20:55 crc kubenswrapper[4861]: I0315 00:20:55.719352 4861 generic.go:334] "Generic (PLEG): container finished" podID="17925b27-9408-4067-970e-de025397edf2" containerID="bb171c29cca24c92ed5c18948fb05536ebeb996d6a9b0f12c0ecea8e09e5d55d" exitCode=0 Mar 15 00:20:55 crc kubenswrapper[4861]: I0315 00:20:55.719468 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerDied","Data":"bb171c29cca24c92ed5c18948fb05536ebeb996d6a9b0f12c0ecea8e09e5d55d"} Mar 15 00:20:55 crc kubenswrapper[4861]: I0315 00:20:55.722606 4861 generic.go:334] "Generic (PLEG): container finished" podID="2c747e57-e783-45d0-a353-521096017d15" containerID="a5b75f50b33c0a58f0229168ab9b8dbd5802b501d896009c7de3a07558f41914" exitCode=0 Mar 15 00:20:55 crc kubenswrapper[4861]: I0315 00:20:55.722706 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" event={"ID":"2c747e57-e783-45d0-a353-521096017d15","Type":"ContainerDied","Data":"a5b75f50b33c0a58f0229168ab9b8dbd5802b501d896009c7de3a07558f41914"} Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.055713 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.073824 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util\") pod \"4e23db65-ec89-427d-aaa4-21d9c00779fd\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.073971 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle\") pod \"4e23db65-ec89-427d-aaa4-21d9c00779fd\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.074051 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfsmx\" (UniqueName: \"kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx\") pod \"4e23db65-ec89-427d-aaa4-21d9c00779fd\" (UID: \"4e23db65-ec89-427d-aaa4-21d9c00779fd\") " Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.076702 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle" (OuterVolumeSpecName: "bundle") pod "4e23db65-ec89-427d-aaa4-21d9c00779fd" (UID: "4e23db65-ec89-427d-aaa4-21d9c00779fd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.084714 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx" (OuterVolumeSpecName: "kube-api-access-kfsmx") pod "4e23db65-ec89-427d-aaa4-21d9c00779fd" (UID: "4e23db65-ec89-427d-aaa4-21d9c00779fd"). InnerVolumeSpecName "kube-api-access-kfsmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.100820 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util" (OuterVolumeSpecName: "util") pod "4e23db65-ec89-427d-aaa4-21d9c00779fd" (UID: "4e23db65-ec89-427d-aaa4-21d9c00779fd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.175402 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfsmx\" (UniqueName: \"kubernetes.io/projected/4e23db65-ec89-427d-aaa4-21d9c00779fd-kube-api-access-kfsmx\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.175435 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.175445 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4e23db65-ec89-427d-aaa4-21d9c00779fd-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.743040 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.743745 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn" event={"ID":"4e23db65-ec89-427d-aaa4-21d9c00779fd","Type":"ContainerDied","Data":"c073354f0b6afa8db93bfcd71d7a215edcce28bcbecedf67e5db7438388e8ce8"} Mar 15 00:20:56 crc kubenswrapper[4861]: I0315 00:20:56.743810 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c073354f0b6afa8db93bfcd71d7a215edcce28bcbecedf67e5db7438388e8ce8" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.000672 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.007776 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.190841 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle\") pod \"17925b27-9408-4067-970e-de025397edf2\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.192317 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh6q8\" (UniqueName: \"kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8\") pod \"2c747e57-e783-45d0-a353-521096017d15\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.191919 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle" (OuterVolumeSpecName: "bundle") pod "17925b27-9408-4067-970e-de025397edf2" (UID: "17925b27-9408-4067-970e-de025397edf2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.192625 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util\") pod \"2c747e57-e783-45d0-a353-521096017d15\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.192991 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util\") pod \"17925b27-9408-4067-970e-de025397edf2\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.194087 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khn84\" (UniqueName: \"kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84\") pod \"17925b27-9408-4067-970e-de025397edf2\" (UID: \"17925b27-9408-4067-970e-de025397edf2\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.194338 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle\") pod \"2c747e57-e783-45d0-a353-521096017d15\" (UID: \"2c747e57-e783-45d0-a353-521096017d15\") " Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.194921 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.195753 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle" (OuterVolumeSpecName: "bundle") pod "2c747e57-e783-45d0-a353-521096017d15" (UID: "2c747e57-e783-45d0-a353-521096017d15"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.197062 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84" (OuterVolumeSpecName: "kube-api-access-khn84") pod "17925b27-9408-4067-970e-de025397edf2" (UID: "17925b27-9408-4067-970e-de025397edf2"). InnerVolumeSpecName "kube-api-access-khn84". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.197629 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8" (OuterVolumeSpecName: "kube-api-access-xh6q8") pod "2c747e57-e783-45d0-a353-521096017d15" (UID: "2c747e57-e783-45d0-a353-521096017d15"). InnerVolumeSpecName "kube-api-access-xh6q8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.204961 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util" (OuterVolumeSpecName: "util") pod "17925b27-9408-4067-970e-de025397edf2" (UID: "17925b27-9408-4067-970e-de025397edf2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.224290 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util" (OuterVolumeSpecName: "util") pod "2c747e57-e783-45d0-a353-521096017d15" (UID: "2c747e57-e783-45d0-a353-521096017d15"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.296257 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.296792 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/17925b27-9408-4067-970e-de025397edf2-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.296966 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khn84\" (UniqueName: \"kubernetes.io/projected/17925b27-9408-4067-970e-de025397edf2-kube-api-access-khn84\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.297127 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c747e57-e783-45d0-a353-521096017d15-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.297347 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh6q8\" (UniqueName: \"kubernetes.io/projected/2c747e57-e783-45d0-a353-521096017d15-kube-api-access-xh6q8\") on node \"crc\" DevicePath \"\"" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.754037 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" event={"ID":"17925b27-9408-4067-970e-de025397edf2","Type":"ContainerDied","Data":"565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3"} Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.754549 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="565898642a97dc22c9c7b66843ed0607c49240c3ed231dbf92b7e7081d1148c3" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.754132 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.759163 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" event={"ID":"2c747e57-e783-45d0-a353-521096017d15","Type":"ContainerDied","Data":"e8c2df49701d2fa7e2aced354300c5c3547b33569c232acd9dfd59fedac19c18"} Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.759224 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c2df49701d2fa7e2aced354300c5c3547b33569c232acd9dfd59fedac19c18" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.759290 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.880920 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr"] Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881165 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881178 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881190 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881195 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881205 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881210 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881219 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881225 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881238 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881245 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881259 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881267 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="util" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881277 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881284 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881294 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881301 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: E0315 00:20:57.881309 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881316 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="pull" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881409 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="17925b27-9408-4067-970e-de025397edf2" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881421 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c747e57-e783-45d0-a353-521096017d15" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.881429 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e23db65-ec89-427d-aaa4-21d9c00779fd" containerName="extract" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.882171 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.887648 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.897061 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr"] Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.907372 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.907507 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:57 crc kubenswrapper[4861]: I0315 00:20:57.907579 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fk89\" (UniqueName: \"kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.009630 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.009789 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.009833 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fk89\" (UniqueName: \"kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.010857 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.011117 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.031359 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fk89\" (UniqueName: \"kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.204303 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.446899 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr"] Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.767990 4861 generic.go:334] "Generic (PLEG): container finished" podID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerID="1800b1826c18e0c88bd9770750bd364717496d3a7c157e55f3a812870cf081cb" exitCode=0 Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.768053 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" event={"ID":"fc5af859-0f30-4bf8-8eb3-8101cee89857","Type":"ContainerDied","Data":"1800b1826c18e0c88bd9770750bd364717496d3a7c157e55f3a812870cf081cb"} Mar 15 00:20:58 crc kubenswrapper[4861]: I0315 00:20:58.768093 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" event={"ID":"fc5af859-0f30-4bf8-8eb3-8101cee89857","Type":"ContainerStarted","Data":"47b54dcfa4536cc8045f35a1f7c9a2f31c461ce3fc970e98eb086dd2899fd8b3"} Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.581103 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-hmrwn"] Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.582489 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.585057 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-qkvhw" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.585134 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.589581 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.630233 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-hmrwn"] Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.664172 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g45k\" (UniqueName: \"kubernetes.io/projected/8dd5e3b8-213b-4f25-bacd-3fbc45147c62-kube-api-access-4g45k\") pod \"interconnect-operator-5bb49f789d-hmrwn\" (UID: \"8dd5e3b8-213b-4f25-bacd-3fbc45147c62\") " pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.766103 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g45k\" (UniqueName: \"kubernetes.io/projected/8dd5e3b8-213b-4f25-bacd-3fbc45147c62-kube-api-access-4g45k\") pod \"interconnect-operator-5bb49f789d-hmrwn\" (UID: \"8dd5e3b8-213b-4f25-bacd-3fbc45147c62\") " pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.804517 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g45k\" (UniqueName: \"kubernetes.io/projected/8dd5e3b8-213b-4f25-bacd-3fbc45147c62-kube-api-access-4g45k\") pod \"interconnect-operator-5bb49f789d-hmrwn\" (UID: \"8dd5e3b8-213b-4f25-bacd-3fbc45147c62\") " pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" Mar 15 00:21:00 crc kubenswrapper[4861]: I0315 00:21:00.900245 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.282534 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.282652 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.282735 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.283663 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.283747 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0" gracePeriod=600 Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.628261 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-hmrwn"] Mar 15 00:21:02 crc kubenswrapper[4861]: W0315 00:21:02.639325 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dd5e3b8_213b_4f25_bacd_3fbc45147c62.slice/crio-696f14c031488db6f1a3f47c9134d0bba523a59fe790983eb68aa8ede78304a1 WatchSource:0}: Error finding container 696f14c031488db6f1a3f47c9134d0bba523a59fe790983eb68aa8ede78304a1: Status 404 returned error can't find the container with id 696f14c031488db6f1a3f47c9134d0bba523a59fe790983eb68aa8ede78304a1 Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.642844 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.800021 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" event={"ID":"8dd5e3b8-213b-4f25-bacd-3fbc45147c62","Type":"ContainerStarted","Data":"696f14c031488db6f1a3f47c9134d0bba523a59fe790983eb68aa8ede78304a1"} Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.803385 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0" exitCode=0 Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.803435 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0"} Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.803514 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0"} Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.803545 4861 scope.go:117] "RemoveContainer" containerID="f1ed7bcaa19390c2028f08cf37a69d438c059ad05285e0219f8bbc39b0570a1e" Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.806412 4861 generic.go:334] "Generic (PLEG): container finished" podID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerID="68ca32b740b0855207ed8c8b532eb45865578dd502f89af41699a78dee0c26dc" exitCode=0 Mar 15 00:21:02 crc kubenswrapper[4861]: I0315 00:21:02.806509 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" event={"ID":"fc5af859-0f30-4bf8-8eb3-8101cee89857","Type":"ContainerDied","Data":"68ca32b740b0855207ed8c8b532eb45865578dd502f89af41699a78dee0c26dc"} Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.841756 4861 generic.go:334] "Generic (PLEG): container finished" podID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerID="4895675b4b0d0ba6ba623dbfe6346042c659a9b4f12449070d9d75d273a16d29" exitCode=0 Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.841920 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" event={"ID":"fc5af859-0f30-4bf8-8eb3-8101cee89857","Type":"ContainerDied","Data":"4895675b4b0d0ba6ba623dbfe6346042c659a9b4f12449070d9d75d273a16d29"} Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.993500 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-74bcbc86d7-q6xmd"] Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.994263 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.996098 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Mar 15 00:21:03 crc kubenswrapper[4861]: I0315 00:21:03.998053 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-mf2l9" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.007104 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-74bcbc86d7-q6xmd"] Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.016946 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-webhook-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.017073 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8whx7\" (UniqueName: \"kubernetes.io/projected/f518207d-da6b-498a-afeb-1a86f16aa5c1-kube-api-access-8whx7\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.017121 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-apiservice-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.123032 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8whx7\" (UniqueName: \"kubernetes.io/projected/f518207d-da6b-498a-afeb-1a86f16aa5c1-kube-api-access-8whx7\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.123123 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-apiservice-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.123192 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-webhook-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.132841 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-apiservice-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.132936 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f518207d-da6b-498a-afeb-1a86f16aa5c1-webhook-cert\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.147511 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8whx7\" (UniqueName: \"kubernetes.io/projected/f518207d-da6b-498a-afeb-1a86f16aa5c1-kube-api-access-8whx7\") pod \"elastic-operator-74bcbc86d7-q6xmd\" (UID: \"f518207d-da6b-498a-afeb-1a86f16aa5c1\") " pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.317476 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.577456 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-74bcbc86d7-q6xmd"] Mar 15 00:21:04 crc kubenswrapper[4861]: W0315 00:21:04.590914 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf518207d_da6b_498a_afeb_1a86f16aa5c1.slice/crio-f89e4c181d9f6aeb97b5073e49bee75627c4609c2b3b4b9134580eace9943807 WatchSource:0}: Error finding container f89e4c181d9f6aeb97b5073e49bee75627c4609c2b3b4b9134580eace9943807: Status 404 returned error can't find the container with id f89e4c181d9f6aeb97b5073e49bee75627c4609c2b3b4b9134580eace9943807 Mar 15 00:21:04 crc kubenswrapper[4861]: I0315 00:21:04.854717 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" event={"ID":"f518207d-da6b-498a-afeb-1a86f16aa5c1","Type":"ContainerStarted","Data":"f89e4c181d9f6aeb97b5073e49bee75627c4609c2b3b4b9134580eace9943807"} Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.134411 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.251522 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fk89\" (UniqueName: \"kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89\") pod \"fc5af859-0f30-4bf8-8eb3-8101cee89857\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.252396 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util\") pod \"fc5af859-0f30-4bf8-8eb3-8101cee89857\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.252551 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle\") pod \"fc5af859-0f30-4bf8-8eb3-8101cee89857\" (UID: \"fc5af859-0f30-4bf8-8eb3-8101cee89857\") " Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.253767 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle" (OuterVolumeSpecName: "bundle") pod "fc5af859-0f30-4bf8-8eb3-8101cee89857" (UID: "fc5af859-0f30-4bf8-8eb3-8101cee89857"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.255649 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.260673 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89" (OuterVolumeSpecName: "kube-api-access-9fk89") pod "fc5af859-0f30-4bf8-8eb3-8101cee89857" (UID: "fc5af859-0f30-4bf8-8eb3-8101cee89857"). InnerVolumeSpecName "kube-api-access-9fk89". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.266667 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util" (OuterVolumeSpecName: "util") pod "fc5af859-0f30-4bf8-8eb3-8101cee89857" (UID: "fc5af859-0f30-4bf8-8eb3-8101cee89857"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.357161 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fk89\" (UniqueName: \"kubernetes.io/projected/fc5af859-0f30-4bf8-8eb3-8101cee89857-kube-api-access-9fk89\") on node \"crc\" DevicePath \"\"" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.357192 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc5af859-0f30-4bf8-8eb3-8101cee89857-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.862543 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" event={"ID":"fc5af859-0f30-4bf8-8eb3-8101cee89857","Type":"ContainerDied","Data":"47b54dcfa4536cc8045f35a1f7c9a2f31c461ce3fc970e98eb086dd2899fd8b3"} Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.862996 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47b54dcfa4536cc8045f35a1f7c9a2f31c461ce3fc970e98eb086dd2899fd8b3" Mar 15 00:21:05 crc kubenswrapper[4861]: I0315 00:21:05.863069 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr" Mar 15 00:21:12 crc kubenswrapper[4861]: I0315 00:21:12.922981 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" event={"ID":"f518207d-da6b-498a-afeb-1a86f16aa5c1","Type":"ContainerStarted","Data":"23e823a39b89c1307de1e4e45f3540b55bec18f7f61b235042b179a056c3227e"} Mar 15 00:21:12 crc kubenswrapper[4861]: I0315 00:21:12.927910 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" event={"ID":"8dd5e3b8-213b-4f25-bacd-3fbc45147c62","Type":"ContainerStarted","Data":"4d10f28773bcaf21e74dbe4e7b466404eccf065cb01a691c7e33a5a4bcf47d93"} Mar 15 00:21:12 crc kubenswrapper[4861]: I0315 00:21:12.956023 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-74bcbc86d7-q6xmd" podStartSLOduration=2.706475217 podStartE2EDuration="9.955998308s" podCreationTimestamp="2026-03-15 00:21:03 +0000 UTC" firstStartedPulling="2026-03-15 00:21:04.593369821 +0000 UTC m=+857.905567354" lastFinishedPulling="2026-03-15 00:21:11.842892912 +0000 UTC m=+865.155090445" observedRunningTime="2026-03-15 00:21:12.944593755 +0000 UTC m=+866.256791308" watchObservedRunningTime="2026-03-15 00:21:12.955998308 +0000 UTC m=+866.268195841" Mar 15 00:21:12 crc kubenswrapper[4861]: I0315 00:21:12.978808 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-hmrwn" podStartSLOduration=3.755002302 podStartE2EDuration="12.978781905s" podCreationTimestamp="2026-03-15 00:21:00 +0000 UTC" firstStartedPulling="2026-03-15 00:21:02.642497292 +0000 UTC m=+855.954694835" lastFinishedPulling="2026-03-15 00:21:11.866276905 +0000 UTC m=+865.178474438" observedRunningTime="2026-03-15 00:21:12.973394877 +0000 UTC m=+866.285592430" watchObservedRunningTime="2026-03-15 00:21:12.978781905 +0000 UTC m=+866.290979438" Mar 15 00:21:14 crc kubenswrapper[4861]: I0315 00:21:14.463681 4861 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.336525 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw"] Mar 15 00:21:18 crc kubenswrapper[4861]: E0315 00:21:18.337489 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="util" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.337508 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="util" Mar 15 00:21:18 crc kubenswrapper[4861]: E0315 00:21:18.337519 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="extract" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.337526 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="extract" Mar 15 00:21:18 crc kubenswrapper[4861]: E0315 00:21:18.337541 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="pull" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.337548 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="pull" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.337666 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc5af859-0f30-4bf8-8eb3-8101cee89857" containerName="extract" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.338135 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.340985 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.341309 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.341511 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-2dssm" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.359664 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.452009 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5kw\" (UniqueName: \"kubernetes.io/projected/57e95c2a-a805-46c7-b30e-5e583a122421-kube-api-access-vn5kw\") pod \"obo-prometheus-operator-68bc856cb9-x6mgw\" (UID: \"57e95c2a-a805-46c7-b30e-5e583a122421\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.467511 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.468337 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.474909 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.474984 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-s6bvm" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.490063 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.495407 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.496250 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.506920 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.553696 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5kw\" (UniqueName: \"kubernetes.io/projected/57e95c2a-a805-46c7-b30e-5e583a122421-kube-api-access-vn5kw\") pod \"obo-prometheus-operator-68bc856cb9-x6mgw\" (UID: \"57e95c2a-a805-46c7-b30e-5e583a122421\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.553844 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.553938 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.590139 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5kw\" (UniqueName: \"kubernetes.io/projected/57e95c2a-a805-46c7-b30e-5e583a122421-kube-api-access-vn5kw\") pod \"obo-prometheus-operator-68bc856cb9-x6mgw\" (UID: \"57e95c2a-a805-46c7-b30e-5e583a122421\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.655369 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.655448 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.655516 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.655572 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.659604 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.659786 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.675185 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/01397ae2-5789-4479-8670-ffea619a6e15-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7\" (UID: \"01397ae2-5789-4479-8670-ffea619a6e15\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.678076 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-j58pz"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.678857 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.682015 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.683129 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dbj6l" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.700965 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-j58pz"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.756874 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6p7k\" (UniqueName: \"kubernetes.io/projected/dee11996-b156-48ac-ac7c-cf3df3bad8e5-kube-api-access-n6p7k\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.756966 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.757200 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dee11996-b156-48ac-ac7c-cf3df3bad8e5-observability-operator-tls\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.757350 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.765329 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.774552 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1dc0d506-b1fd-4326-9417-67d37744ed0b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7\" (UID: \"1dc0d506-b1fd-4326-9417-67d37744ed0b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.782448 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.812481 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.863531 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6p7k\" (UniqueName: \"kubernetes.io/projected/dee11996-b156-48ac-ac7c-cf3df3bad8e5-kube-api-access-n6p7k\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.863690 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dee11996-b156-48ac-ac7c-cf3df3bad8e5-observability-operator-tls\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.866439 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-qr9rk"] Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.869114 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dee11996-b156-48ac-ac7c-cf3df3bad8e5-observability-operator-tls\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.875236 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.878828 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6mkld" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.914205 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6p7k\" (UniqueName: \"kubernetes.io/projected/dee11996-b156-48ac-ac7c-cf3df3bad8e5-kube-api-access-n6p7k\") pod \"observability-operator-59bdc8b94-j58pz\" (UID: \"dee11996-b156-48ac-ac7c-cf3df3bad8e5\") " pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:18 crc kubenswrapper[4861]: I0315 00:21:18.945882 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-qr9rk"] Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.025076 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.090093 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/41a1f7ae-f37f-4d4f-a12b-927e525730c8-openshift-service-ca\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.090160 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg6lx\" (UniqueName: \"kubernetes.io/projected/41a1f7ae-f37f-4d4f-a12b-927e525730c8-kube-api-access-hg6lx\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.193339 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/41a1f7ae-f37f-4d4f-a12b-927e525730c8-openshift-service-ca\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.193412 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg6lx\" (UniqueName: \"kubernetes.io/projected/41a1f7ae-f37f-4d4f-a12b-927e525730c8-kube-api-access-hg6lx\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.194827 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/41a1f7ae-f37f-4d4f-a12b-927e525730c8-openshift-service-ca\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.195819 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw"] Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.218516 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg6lx\" (UniqueName: \"kubernetes.io/projected/41a1f7ae-f37f-4d4f-a12b-927e525730c8-kube-api-access-hg6lx\") pod \"perses-operator-5bf474d74f-qr9rk\" (UID: \"41a1f7ae-f37f-4d4f-a12b-927e525730c8\") " pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.262298 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7"] Mar 15 00:21:19 crc kubenswrapper[4861]: W0315 00:21:19.275719 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1dc0d506_b1fd_4326_9417_67d37744ed0b.slice/crio-8288cab25ddbb255dc5661fa1ca7434c339a88c3b108cc2ce8461add2c19a3b4 WatchSource:0}: Error finding container 8288cab25ddbb255dc5661fa1ca7434c339a88c3b108cc2ce8461add2c19a3b4: Status 404 returned error can't find the container with id 8288cab25ddbb255dc5661fa1ca7434c339a88c3b108cc2ce8461add2c19a3b4 Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.398292 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" event={"ID":"1dc0d506-b1fd-4326-9417-67d37744ed0b","Type":"ContainerStarted","Data":"8288cab25ddbb255dc5661fa1ca7434c339a88c3b108cc2ce8461add2c19a3b4"} Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.432530 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" event={"ID":"57e95c2a-a805-46c7-b30e-5e583a122421","Type":"ContainerStarted","Data":"31c44e0845e2fbacf711ff3f361982a8f88ce45c854d4258c791bec778ef1e81"} Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.432594 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7"] Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.510717 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.584786 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-j58pz"] Mar 15 00:21:19 crc kubenswrapper[4861]: I0315 00:21:19.908355 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-qr9rk"] Mar 15 00:21:20 crc kubenswrapper[4861]: I0315 00:21:20.481796 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" event={"ID":"41a1f7ae-f37f-4d4f-a12b-927e525730c8","Type":"ContainerStarted","Data":"6eb2f441403fab323adda6a99faab521dc7a63ed0773297a5ba20c68fdb38c4a"} Mar 15 00:21:20 crc kubenswrapper[4861]: I0315 00:21:20.503683 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" event={"ID":"01397ae2-5789-4479-8670-ffea619a6e15","Type":"ContainerStarted","Data":"5f4326b1297ea1670a79a7a0c0e7c480632af11eb458762e515995270a35e2b3"} Mar 15 00:21:20 crc kubenswrapper[4861]: I0315 00:21:20.511237 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" event={"ID":"dee11996-b156-48ac-ac7c-cf3df3bad8e5","Type":"ContainerStarted","Data":"474a13487dacfd83cc9274f2e87123235aa49a399fd818805c45abddb75fdd69"} Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.544515 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.551886 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.559612 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.559797 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.559814 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-xmcwh" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.559879 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.560060 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.560189 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.568647 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.569752 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.569825 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.592974 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649628 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649683 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649716 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/47c92463-1387-4e36-a0a4-00ced68121fd-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649738 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649759 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649775 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649794 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649821 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649851 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649872 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649892 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649911 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649939 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649959 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.649983 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751356 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751406 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751442 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/47c92463-1387-4e36-a0a4-00ced68121fd-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751459 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751488 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751508 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751535 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751577 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751599 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751620 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751641 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751661 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751682 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751701 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.751724 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.752018 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.758833 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.760541 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.760745 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.761135 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.761138 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.778871 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/47c92463-1387-4e36-a0a4-00ced68121fd-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.780196 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.781040 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.788342 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.788890 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.790016 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.790202 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/47c92463-1387-4e36-a0a4-00ced68121fd-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.790704 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.791742 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/47c92463-1387-4e36-a0a4-00ced68121fd-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"47c92463-1387-4e36-a0a4-00ced68121fd\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:21 crc kubenswrapper[4861]: I0315 00:21:21.882192 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:21:22 crc kubenswrapper[4861]: I0315 00:21:22.387816 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 15 00:21:22 crc kubenswrapper[4861]: I0315 00:21:22.542092 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"47c92463-1387-4e36-a0a4-00ced68121fd","Type":"ContainerStarted","Data":"43657a5cc29c805258830ad4c9f62fa5cff5488d56d03d3067d7a611eb3286df"} Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.295506 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv"] Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.296719 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.301009 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.301790 4861 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-7svnd" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.302397 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.313631 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv"] Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.321063 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z58v5\" (UniqueName: \"kubernetes.io/projected/05a4b85f-5129-40fd-81ef-f0d876e09dbb-kube-api-access-z58v5\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.321138 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/05a4b85f-5129-40fd-81ef-f0d876e09dbb-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.422287 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/05a4b85f-5129-40fd-81ef-f0d876e09dbb-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.422354 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z58v5\" (UniqueName: \"kubernetes.io/projected/05a4b85f-5129-40fd-81ef-f0d876e09dbb-kube-api-access-z58v5\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.423065 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/05a4b85f-5129-40fd-81ef-f0d876e09dbb-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.444302 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z58v5\" (UniqueName: \"kubernetes.io/projected/05a4b85f-5129-40fd-81ef-f0d876e09dbb-kube-api-access-z58v5\") pod \"cert-manager-operator-controller-manager-5586865c96-hc2zv\" (UID: \"05a4b85f-5129-40fd-81ef-f0d876e09dbb\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:25 crc kubenswrapper[4861]: I0315 00:21:25.625403 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" Mar 15 00:21:35 crc kubenswrapper[4861]: I0315 00:21:35.672975 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv"] Mar 15 00:21:47 crc kubenswrapper[4861]: I0315 00:21:47.736568 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" event={"ID":"41a1f7ae-f37f-4d4f-a12b-927e525730c8","Type":"ContainerStarted","Data":"fd77b839c0feed1c8c24cb1bae44c47c66282840ea2e27a3319a5f294cb26626"} Mar 15 00:21:47 crc kubenswrapper[4861]: I0315 00:21:47.737709 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:47 crc kubenswrapper[4861]: I0315 00:21:47.740636 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" event={"ID":"05a4b85f-5129-40fd-81ef-f0d876e09dbb","Type":"ContainerStarted","Data":"90946694a85bd17421a94d21bde92425fa8f5c913debe5752a44c69f8352718a"} Mar 15 00:21:47 crc kubenswrapper[4861]: E0315 00:21:47.862280 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Mar 15 00:21:47 crc kubenswrapper[4861]: E0315 00:21:47.862698 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(47c92463-1387-4e36-a0a4-00ced68121fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 15 00:21:47 crc kubenswrapper[4861]: E0315 00:21:47.864005 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="47c92463-1387-4e36-a0a4-00ced68121fd" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.771769 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" event={"ID":"01397ae2-5789-4479-8670-ffea619a6e15","Type":"ContainerStarted","Data":"297de658e01fbcee906bc669a97539abd86d12998f5d48d802bfad4aebb7af45"} Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.777928 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" event={"ID":"dee11996-b156-48ac-ac7c-cf3df3bad8e5","Type":"ContainerStarted","Data":"ef50debba352b8f52a38dcaf9548ef63c1d61b263e7be2fcc37d07c48dc44a7f"} Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.781825 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.789749 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" event={"ID":"1dc0d506-b1fd-4326-9417-67d37744ed0b","Type":"ContainerStarted","Data":"d0ff159a2ecac9ddfe6850d10d7ddb9073116643a9e1ac96505dcafe9c476f76"} Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.793247 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" event={"ID":"57e95c2a-a805-46c7-b30e-5e583a122421","Type":"ContainerStarted","Data":"e0c5442c7533af6b81946c481a8906d5202c1b359cd951e3e4ba2ec94c07981c"} Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.795364 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7" podStartSLOduration=3.16784853 podStartE2EDuration="30.795339347s" podCreationTimestamp="2026-03-15 00:21:18 +0000 UTC" firstStartedPulling="2026-03-15 00:21:19.453894257 +0000 UTC m=+872.766091790" lastFinishedPulling="2026-03-15 00:21:47.081385074 +0000 UTC m=+900.393582607" observedRunningTime="2026-03-15 00:21:48.794401951 +0000 UTC m=+902.106599504" watchObservedRunningTime="2026-03-15 00:21:48.795339347 +0000 UTC m=+902.107536880" Mar 15 00:21:48 crc kubenswrapper[4861]: E0315 00:21:48.799426 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="47c92463-1387-4e36-a0a4-00ced68121fd" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.817832 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" podStartSLOduration=15.545653577 podStartE2EDuration="30.817802065s" podCreationTimestamp="2026-03-15 00:21:18 +0000 UTC" firstStartedPulling="2026-03-15 00:21:19.917485088 +0000 UTC m=+873.229682621" lastFinishedPulling="2026-03-15 00:21:35.189633576 +0000 UTC m=+888.501831109" observedRunningTime="2026-03-15 00:21:47.760346739 +0000 UTC m=+901.072544262" watchObservedRunningTime="2026-03-15 00:21:48.817802065 +0000 UTC m=+902.129999598" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.837943 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7" podStartSLOduration=4.016699748 podStartE2EDuration="30.837912838s" podCreationTimestamp="2026-03-15 00:21:18 +0000 UTC" firstStartedPulling="2026-03-15 00:21:19.288814826 +0000 UTC m=+872.601012359" lastFinishedPulling="2026-03-15 00:21:46.110027916 +0000 UTC m=+899.422225449" observedRunningTime="2026-03-15 00:21:48.822031141 +0000 UTC m=+902.134228684" watchObservedRunningTime="2026-03-15 00:21:48.837912838 +0000 UTC m=+902.150110371" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.854113 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" podStartSLOduration=4.353552893 podStartE2EDuration="30.854073643s" podCreationTimestamp="2026-03-15 00:21:18 +0000 UTC" firstStartedPulling="2026-03-15 00:21:19.609519497 +0000 UTC m=+872.921717030" lastFinishedPulling="2026-03-15 00:21:46.110040247 +0000 UTC m=+899.422237780" observedRunningTime="2026-03-15 00:21:48.853866427 +0000 UTC m=+902.166063960" watchObservedRunningTime="2026-03-15 00:21:48.854073643 +0000 UTC m=+902.166271176" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.947347 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" Mar 15 00:21:48 crc kubenswrapper[4861]: I0315 00:21:48.969874 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-x6mgw" podStartSLOduration=14.998071924 podStartE2EDuration="30.969846107s" podCreationTimestamp="2026-03-15 00:21:18 +0000 UTC" firstStartedPulling="2026-03-15 00:21:19.217760951 +0000 UTC m=+872.529958484" lastFinishedPulling="2026-03-15 00:21:35.189535134 +0000 UTC m=+888.501732667" observedRunningTime="2026-03-15 00:21:48.918162086 +0000 UTC m=+902.230359619" watchObservedRunningTime="2026-03-15 00:21:48.969846107 +0000 UTC m=+902.282043640" Mar 15 00:21:49 crc kubenswrapper[4861]: I0315 00:21:49.043486 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 15 00:21:49 crc kubenswrapper[4861]: I0315 00:21:49.092944 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 15 00:21:49 crc kubenswrapper[4861]: E0315 00:21:49.806925 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="47c92463-1387-4e36-a0a4-00ced68121fd" Mar 15 00:21:50 crc kubenswrapper[4861]: I0315 00:21:50.811377 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" event={"ID":"05a4b85f-5129-40fd-81ef-f0d876e09dbb","Type":"ContainerStarted","Data":"1bba48757e59d2a51a6a7908abaf2fa57aa710c5449112e374a9eece716e1d8c"} Mar 15 00:21:50 crc kubenswrapper[4861]: E0315 00:21:50.814284 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="47c92463-1387-4e36-a0a4-00ced68121fd" Mar 15 00:21:50 crc kubenswrapper[4861]: I0315 00:21:50.837677 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-hc2zv" podStartSLOduration=22.576306528 podStartE2EDuration="25.837644482s" podCreationTimestamp="2026-03-15 00:21:25 +0000 UTC" firstStartedPulling="2026-03-15 00:21:47.092896661 +0000 UTC m=+900.405094194" lastFinishedPulling="2026-03-15 00:21:50.354234615 +0000 UTC m=+903.666432148" observedRunningTime="2026-03-15 00:21:50.834120775 +0000 UTC m=+904.146318308" watchObservedRunningTime="2026-03-15 00:21:50.837644482 +0000 UTC m=+904.149842015" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.002804 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-v2qct"] Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.005154 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.007845 4861 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-7rvhq" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.012445 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.012908 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.017757 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-v2qct"] Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.165740 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.166285 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgvpv\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-kube-api-access-lgvpv\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.268044 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgvpv\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-kube-api-access-lgvpv\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.268212 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.290936 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgvpv\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-kube-api-access-lgvpv\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.294701 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85ede584-f9ce-419c-8807-f415ecb45d4c-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-v2qct\" (UID: \"85ede584-f9ce-419c-8807-f415ecb45d4c\") " pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.322843 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.614379 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-v2qct"] Mar 15 00:21:53 crc kubenswrapper[4861]: I0315 00:21:53.838393 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" event={"ID":"85ede584-f9ce-419c-8807-f415ecb45d4c","Type":"ContainerStarted","Data":"04f2232796d43ce69d770a36167c941b25679d104e28f3861f8f4fb5a0563863"} Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.025843 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-g5bzq"] Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.026742 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.029635 4861 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-2p5hs" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.034954 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-g5bzq"] Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.195517 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87jv\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-kube-api-access-q87jv\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.195625 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.297252 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87jv\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-kube-api-access-q87jv\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.297350 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.326743 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87jv\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-kube-api-access-q87jv\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.332092 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/febd875f-8a0e-43ab-a41c-d555e44bb04d-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-g5bzq\" (UID: \"febd875f-8a0e-43ab-a41c-d555e44bb04d\") " pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.344431 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.650083 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-g5bzq"] Mar 15 00:21:55 crc kubenswrapper[4861]: W0315 00:21:55.652159 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfebd875f_8a0e_43ab_a41c_d555e44bb04d.slice/crio-3e7c916eaeb760461ea17e6b9d94490627cbd80b2daf3e279022aa0e430f7209 WatchSource:0}: Error finding container 3e7c916eaeb760461ea17e6b9d94490627cbd80b2daf3e279022aa0e430f7209: Status 404 returned error can't find the container with id 3e7c916eaeb760461ea17e6b9d94490627cbd80b2daf3e279022aa0e430f7209 Mar 15 00:21:55 crc kubenswrapper[4861]: I0315 00:21:55.851896 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" event={"ID":"febd875f-8a0e-43ab-a41c-d555e44bb04d","Type":"ContainerStarted","Data":"3e7c916eaeb760461ea17e6b9d94490627cbd80b2daf3e279022aa0e430f7209"} Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.077720 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.084207 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.087245 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.087452 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.087686 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.087872 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.100104 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.157983 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158042 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158073 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158203 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158279 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158339 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158502 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158647 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158685 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158737 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158808 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.158844 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7dck\" (UniqueName: \"kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260159 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7dck\" (UniqueName: \"kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260727 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260759 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260782 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260810 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260832 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260854 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260883 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260922 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260942 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260969 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.260988 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261429 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261489 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261530 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261541 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261544 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261596 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261820 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.261974 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.262437 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.266976 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.275796 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7dck\" (UniqueName: \"kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.291079 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.476569 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.514669 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-qr9rk" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.901914 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" event={"ID":"febd875f-8a0e-43ab-a41c-d555e44bb04d","Type":"ContainerStarted","Data":"7e8d955938d19b23357dc1d57610d412a61f6c63c37e0f329794f57973bcefda"} Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.903865 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" event={"ID":"85ede584-f9ce-419c-8807-f415ecb45d4c","Type":"ContainerStarted","Data":"b97620350042fc12ad95564b079b11ef994b230130d387a17e510d8ae5aa3cb6"} Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.903971 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.918485 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-g5bzq" podStartSLOduration=1.556466981 podStartE2EDuration="4.918466594s" podCreationTimestamp="2026-03-15 00:21:55 +0000 UTC" firstStartedPulling="2026-03-15 00:21:55.656922939 +0000 UTC m=+908.969120472" lastFinishedPulling="2026-03-15 00:21:59.018922552 +0000 UTC m=+912.331120085" observedRunningTime="2026-03-15 00:21:59.917405905 +0000 UTC m=+913.229603458" watchObservedRunningTime="2026-03-15 00:21:59.918466594 +0000 UTC m=+913.230664127" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.966228 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" podStartSLOduration=2.564448479 podStartE2EDuration="7.966201837s" podCreationTimestamp="2026-03-15 00:21:52 +0000 UTC" firstStartedPulling="2026-03-15 00:21:53.623463167 +0000 UTC m=+906.935660700" lastFinishedPulling="2026-03-15 00:21:59.025216525 +0000 UTC m=+912.337414058" observedRunningTime="2026-03-15 00:21:59.960907412 +0000 UTC m=+913.273104945" watchObservedRunningTime="2026-03-15 00:21:59.966201837 +0000 UTC m=+913.278399390" Mar 15 00:21:59 crc kubenswrapper[4861]: I0315 00:21:59.988927 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:21:59 crc kubenswrapper[4861]: W0315 00:21:59.994534 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94246c7e_a444_40fd_b504_f71b327d289f.slice/crio-09b760d5bab8449c47f9fc9793e7ad6bebf0a1125d815a866c24b724c16faad3 WatchSource:0}: Error finding container 09b760d5bab8449c47f9fc9793e7ad6bebf0a1125d815a866c24b724c16faad3: Status 404 returned error can't find the container with id 09b760d5bab8449c47f9fc9793e7ad6bebf0a1125d815a866c24b724c16faad3 Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.124987 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558902-46bp7"] Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.126218 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.128162 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.128227 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.131076 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.134533 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558902-46bp7"] Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.276688 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppdpc\" (UniqueName: \"kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc\") pod \"auto-csr-approver-29558902-46bp7\" (UID: \"5b38eb7a-7876-433d-b110-7030850e9252\") " pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.378682 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppdpc\" (UniqueName: \"kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc\") pod \"auto-csr-approver-29558902-46bp7\" (UID: \"5b38eb7a-7876-433d-b110-7030850e9252\") " pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.400138 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppdpc\" (UniqueName: \"kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc\") pod \"auto-csr-approver-29558902-46bp7\" (UID: \"5b38eb7a-7876-433d-b110-7030850e9252\") " pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.442212 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.682920 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558902-46bp7"] Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.912030 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558902-46bp7" event={"ID":"5b38eb7a-7876-433d-b110-7030850e9252","Type":"ContainerStarted","Data":"b1d6f21cb9f325c579d8b72b2ae722970c30e0224e8ff447fb39bc51d49c2e44"} Mar 15 00:22:00 crc kubenswrapper[4861]: I0315 00:22:00.914195 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"94246c7e-a444-40fd-b504-f71b327d289f","Type":"ContainerStarted","Data":"09b760d5bab8449c47f9fc9793e7ad6bebf0a1125d815a866c24b724c16faad3"} Mar 15 00:22:02 crc kubenswrapper[4861]: I0315 00:22:02.941073 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558902-46bp7" event={"ID":"5b38eb7a-7876-433d-b110-7030850e9252","Type":"ContainerStarted","Data":"0f5eed031a8d9e493499f3b34294fbd558f33718d3293b45ca5e72747cc503a2"} Mar 15 00:22:02 crc kubenswrapper[4861]: I0315 00:22:02.976496 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558902-46bp7" podStartSLOduration=1.874224498 podStartE2EDuration="2.976476206s" podCreationTimestamp="2026-03-15 00:22:00 +0000 UTC" firstStartedPulling="2026-03-15 00:22:00.697364489 +0000 UTC m=+914.009562022" lastFinishedPulling="2026-03-15 00:22:01.799616197 +0000 UTC m=+915.111813730" observedRunningTime="2026-03-15 00:22:02.971342156 +0000 UTC m=+916.283539689" watchObservedRunningTime="2026-03-15 00:22:02.976476206 +0000 UTC m=+916.288673739" Mar 15 00:22:03 crc kubenswrapper[4861]: I0315 00:22:03.949698 4861 generic.go:334] "Generic (PLEG): container finished" podID="5b38eb7a-7876-433d-b110-7030850e9252" containerID="0f5eed031a8d9e493499f3b34294fbd558f33718d3293b45ca5e72747cc503a2" exitCode=0 Mar 15 00:22:03 crc kubenswrapper[4861]: I0315 00:22:03.949750 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558902-46bp7" event={"ID":"5b38eb7a-7876-433d-b110-7030850e9252","Type":"ContainerDied","Data":"0f5eed031a8d9e493499f3b34294fbd558f33718d3293b45ca5e72747cc503a2"} Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.066475 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-d8l4v"] Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.068844 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.077244 4861 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-4khkw" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.093751 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-d8l4v"] Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.142445 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9v66\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-kube-api-access-m9v66\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.142502 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-bound-sa-token\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.244026 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9v66\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-kube-api-access-m9v66\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.244101 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-bound-sa-token\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.289300 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-bound-sa-token\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.289696 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9v66\" (UniqueName: \"kubernetes.io/projected/8197885f-6fb5-4b68-9d32-1c39c74d220f-kube-api-access-m9v66\") pod \"cert-manager-545d4d4674-d8l4v\" (UID: \"8197885f-6fb5-4b68-9d32-1c39c74d220f\") " pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.390952 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-d8l4v" Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.758418 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-d8l4v"] Mar 15 00:22:04 crc kubenswrapper[4861]: I0315 00:22:04.959398 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"47c92463-1387-4e36-a0a4-00ced68121fd","Type":"ContainerStarted","Data":"3d70e5307ef5d0a2d572ad0a0cfd279535436e25706ad350f44b457de448ac36"} Mar 15 00:22:05 crc kubenswrapper[4861]: I0315 00:22:05.971944 4861 generic.go:334] "Generic (PLEG): container finished" podID="47c92463-1387-4e36-a0a4-00ced68121fd" containerID="3d70e5307ef5d0a2d572ad0a0cfd279535436e25706ad350f44b457de448ac36" exitCode=0 Mar 15 00:22:05 crc kubenswrapper[4861]: I0315 00:22:05.971997 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"47c92463-1387-4e36-a0a4-00ced68121fd","Type":"ContainerDied","Data":"3d70e5307ef5d0a2d572ad0a0cfd279535436e25706ad350f44b457de448ac36"} Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.326820 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-v2qct" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.610470 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.716816 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppdpc\" (UniqueName: \"kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc\") pod \"5b38eb7a-7876-433d-b110-7030850e9252\" (UID: \"5b38eb7a-7876-433d-b110-7030850e9252\") " Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.724029 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc" (OuterVolumeSpecName: "kube-api-access-ppdpc") pod "5b38eb7a-7876-433d-b110-7030850e9252" (UID: "5b38eb7a-7876-433d-b110-7030850e9252"). InnerVolumeSpecName "kube-api-access-ppdpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.818661 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppdpc\" (UniqueName: \"kubernetes.io/projected/5b38eb7a-7876-433d-b110-7030850e9252-kube-api-access-ppdpc\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.991822 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558902-46bp7" event={"ID":"5b38eb7a-7876-433d-b110-7030850e9252","Type":"ContainerDied","Data":"b1d6f21cb9f325c579d8b72b2ae722970c30e0224e8ff447fb39bc51d49c2e44"} Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.991850 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558902-46bp7" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.991873 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1d6f21cb9f325c579d8b72b2ae722970c30e0224e8ff447fb39bc51d49c2e44" Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.994353 4861 generic.go:334] "Generic (PLEG): container finished" podID="47c92463-1387-4e36-a0a4-00ced68121fd" containerID="f03972fa374e29e38e7bc4a02e543a1b5e7adfb2a2a1d19d9592ff1ae1579867" exitCode=0 Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.994417 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"47c92463-1387-4e36-a0a4-00ced68121fd","Type":"ContainerDied","Data":"f03972fa374e29e38e7bc4a02e543a1b5e7adfb2a2a1d19d9592ff1ae1579867"} Mar 15 00:22:08 crc kubenswrapper[4861]: I0315 00:22:08.997212 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"94246c7e-a444-40fd-b504-f71b327d289f","Type":"ContainerStarted","Data":"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554"} Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.000056 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-d8l4v" event={"ID":"8197885f-6fb5-4b68-9d32-1c39c74d220f","Type":"ContainerStarted","Data":"e2cfbe3dfe216514b40e0e67c0d39cb7415e1b07d6b47d2e129c54acdfa04e7f"} Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.000096 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-d8l4v" event={"ID":"8197885f-6fb5-4b68-9d32-1c39c74d220f","Type":"ContainerStarted","Data":"9743047a144ec4117d0cb64228e3cc42918076661b7f1abdf8632b28da8620c7"} Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.070073 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-d8l4v" podStartSLOduration=5.0700475130000004 podStartE2EDuration="5.070047513s" podCreationTimestamp="2026-03-15 00:22:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:22:09.064634945 +0000 UTC m=+922.376832498" watchObservedRunningTime="2026-03-15 00:22:09.070047513 +0000 UTC m=+922.382245046" Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.082091 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.685220 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558896-8qxmj"] Mar 15 00:22:09 crc kubenswrapper[4861]: I0315 00:22:09.690244 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558896-8qxmj"] Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.009741 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"47c92463-1387-4e36-a0a4-00ced68121fd","Type":"ContainerStarted","Data":"5b689cbd1ce2b0fe49678b6acec89b400445583974b19cf3e67e80d845e82e1c"} Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.010993 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.051997 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=7.45334145 podStartE2EDuration="49.051977152s" podCreationTimestamp="2026-03-15 00:21:21 +0000 UTC" firstStartedPulling="2026-03-15 00:21:22.411695132 +0000 UTC m=+875.723892655" lastFinishedPulling="2026-03-15 00:22:04.010330824 +0000 UTC m=+917.322528357" observedRunningTime="2026-03-15 00:22:10.050011227 +0000 UTC m=+923.362208760" watchObservedRunningTime="2026-03-15 00:22:10.051977152 +0000 UTC m=+923.364174685" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.721859 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 15 00:22:10 crc kubenswrapper[4861]: E0315 00:22:10.722227 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b38eb7a-7876-433d-b110-7030850e9252" containerName="oc" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.722248 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b38eb7a-7876-433d-b110-7030850e9252" containerName="oc" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.722441 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b38eb7a-7876-433d-b110-7030850e9252" containerName="oc" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.723935 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.726792 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.728007 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.740190 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.745032 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.866678 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.866734 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.866822 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.866876 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.866972 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867039 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867088 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867120 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867234 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvsqb\" (UniqueName: \"kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867337 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867499 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.867576 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968809 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968867 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968905 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968932 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968953 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.968992 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969162 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969035 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969234 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969289 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvsqb\" (UniqueName: \"kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969279 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969475 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969508 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969530 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.969728 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.970043 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.970061 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.970261 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.970425 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.970686 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.971373 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.977060 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.977398 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:10 crc kubenswrapper[4861]: I0315 00:22:10.994988 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvsqb\" (UniqueName: \"kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb\") pod \"service-telemetry-operator-2-build\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.018742 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="94246c7e-a444-40fd-b504-f71b327d289f" containerName="manage-dockerfile" containerID="cri-o://642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554" gracePeriod=30 Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.043188 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.389963 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Mar 15 00:22:11 crc kubenswrapper[4861]: W0315 00:22:11.398435 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14955a1e_daa9_4c2c_962f_acd6eb7079ae.slice/crio-d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e WatchSource:0}: Error finding container d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e: Status 404 returned error can't find the container with id d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.423734 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f75e814-555e-4ac9-ac2e-db0d34c970fc" path="/var/lib/kubelet/pods/7f75e814-555e-4ac9-ac2e-db0d34c970fc/volumes" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.505623 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_94246c7e-a444-40fd-b504-f71b327d289f/manage-dockerfile/0.log" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.505763 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580297 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7dck\" (UniqueName: \"kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580400 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580471 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580498 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580537 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580616 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580637 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580701 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580736 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580766 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580813 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.580829 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs\") pod \"94246c7e-a444-40fd-b504-f71b327d289f\" (UID: \"94246c7e-a444-40fd-b504-f71b327d289f\") " Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.583087 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.583585 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.583624 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.583923 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.583953 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.584001 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.584158 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.584230 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.584765 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.591126 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.605262 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck" (OuterVolumeSpecName: "kube-api-access-f7dck") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "kube-api-access-f7dck". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.628809 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "94246c7e-a444-40fd-b504-f71b327d289f" (UID: "94246c7e-a444-40fd-b504-f71b327d289f"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686244 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686294 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686308 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/94246c7e-a444-40fd-b504-f71b327d289f-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686318 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686327 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686339 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7dck\" (UniqueName: \"kubernetes.io/projected/94246c7e-a444-40fd-b504-f71b327d289f-kube-api-access-f7dck\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686350 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686359 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/94246c7e-a444-40fd-b504-f71b327d289f-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686369 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686379 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686389 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/94246c7e-a444-40fd-b504-f71b327d289f-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:11 crc kubenswrapper[4861]: I0315 00:22:11.686398 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/94246c7e-a444-40fd-b504-f71b327d289f-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.026631 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerStarted","Data":"51d100be965aa927eda652833174e4b7a1958cdc3ce3dabe1e5c4b142f2d6aa4"} Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.026690 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerStarted","Data":"d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e"} Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030195 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_94246c7e-a444-40fd-b504-f71b327d289f/manage-dockerfile/0.log" Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030237 4861 generic.go:334] "Generic (PLEG): container finished" podID="94246c7e-a444-40fd-b504-f71b327d289f" containerID="642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554" exitCode=1 Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030265 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"94246c7e-a444-40fd-b504-f71b327d289f","Type":"ContainerDied","Data":"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554"} Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030283 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"94246c7e-a444-40fd-b504-f71b327d289f","Type":"ContainerDied","Data":"09b760d5bab8449c47f9fc9793e7ad6bebf0a1125d815a866c24b724c16faad3"} Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030304 4861 scope.go:117] "RemoveContainer" containerID="642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554" Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.030417 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.108542 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.120028 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.144676 4861 scope.go:117] "RemoveContainer" containerID="642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554" Mar 15 00:22:12 crc kubenswrapper[4861]: E0315 00:22:12.145277 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554\": container with ID starting with 642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554 not found: ID does not exist" containerID="642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554" Mar 15 00:22:12 crc kubenswrapper[4861]: I0315 00:22:12.145315 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554"} err="failed to get container status \"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554\": rpc error: code = NotFound desc = could not find container \"642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554\": container with ID starting with 642673682e93df1db654fc66618d20a5be532e05982bb9a5dccbdaef3c77d554 not found: ID does not exist" Mar 15 00:22:13 crc kubenswrapper[4861]: I0315 00:22:13.417152 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94246c7e-a444-40fd-b504-f71b327d289f" path="/var/lib/kubelet/pods/94246c7e-a444-40fd-b504-f71b327d289f/volumes" Mar 15 00:22:19 crc kubenswrapper[4861]: I0315 00:22:19.079358 4861 generic.go:334] "Generic (PLEG): container finished" podID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerID="51d100be965aa927eda652833174e4b7a1958cdc3ce3dabe1e5c4b142f2d6aa4" exitCode=0 Mar 15 00:22:19 crc kubenswrapper[4861]: I0315 00:22:19.079493 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerDied","Data":"51d100be965aa927eda652833174e4b7a1958cdc3ce3dabe1e5c4b142f2d6aa4"} Mar 15 00:22:20 crc kubenswrapper[4861]: I0315 00:22:20.089242 4861 generic.go:334] "Generic (PLEG): container finished" podID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerID="d6bab504d0a39a639bd81069135f2e75f9013d04e22b9b262d3785ddb9403d72" exitCode=0 Mar 15 00:22:20 crc kubenswrapper[4861]: I0315 00:22:20.089306 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerDied","Data":"d6bab504d0a39a639bd81069135f2e75f9013d04e22b9b262d3785ddb9403d72"} Mar 15 00:22:20 crc kubenswrapper[4861]: I0315 00:22:20.143040 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_14955a1e-daa9-4c2c-962f-acd6eb7079ae/manage-dockerfile/0.log" Mar 15 00:22:21 crc kubenswrapper[4861]: I0315 00:22:21.107363 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerStarted","Data":"43d26e8759b6b82a7a278f2e4192f712f66101fb3e2b1a8809878a734135c75b"} Mar 15 00:22:21 crc kubenswrapper[4861]: I0315 00:22:21.148862 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=11.148832836 podStartE2EDuration="11.148832836s" podCreationTimestamp="2026-03-15 00:22:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:22:21.139374246 +0000 UTC m=+934.451571839" watchObservedRunningTime="2026-03-15 00:22:21.148832836 +0000 UTC m=+934.461030399" Mar 15 00:22:22 crc kubenswrapper[4861]: I0315 00:22:22.048744 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="47c92463-1387-4e36-a0a4-00ced68121fd" containerName="elasticsearch" probeResult="failure" output=< Mar 15 00:22:22 crc kubenswrapper[4861]: {"timestamp": "2026-03-15T00:22:22+00:00", "message": "readiness probe failed", "curl_rc": "7"} Mar 15 00:22:22 crc kubenswrapper[4861]: > Mar 15 00:22:27 crc kubenswrapper[4861]: I0315 00:22:27.398201 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 15 00:22:48 crc kubenswrapper[4861]: I0315 00:22:48.075962 4861 scope.go:117] "RemoveContainer" containerID="de14f8b2aee9969d453d844e0e5179d2b0bbb786caf1f0d2af429ebb9853091e" Mar 15 00:23:02 crc kubenswrapper[4861]: I0315 00:23:02.281793 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:23:02 crc kubenswrapper[4861]: I0315 00:23:02.282573 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:23:20 crc kubenswrapper[4861]: I0315 00:23:20.068418 4861 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-j58pz container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.48:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:23:20 crc kubenswrapper[4861]: I0315 00:23:20.068446 4861 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-j58pz container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.48:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 15 00:23:20 crc kubenswrapper[4861]: I0315 00:23:20.069249 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" podUID="dee11996-b156-48ac-ac7c-cf3df3bad8e5" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.48:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 15 00:23:20 crc kubenswrapper[4861]: I0315 00:23:20.069292 4861 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-j58pz" podUID="dee11996-b156-48ac-ac7c-cf3df3bad8e5" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.48:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 15 00:23:32 crc kubenswrapper[4861]: I0315 00:23:32.282006 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:23:32 crc kubenswrapper[4861]: I0315 00:23:32.282588 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:23:51 crc kubenswrapper[4861]: I0315 00:23:51.164380 4861 generic.go:334] "Generic (PLEG): container finished" podID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerID="43d26e8759b6b82a7a278f2e4192f712f66101fb3e2b1a8809878a734135c75b" exitCode=0 Mar 15 00:23:51 crc kubenswrapper[4861]: I0315 00:23:51.164519 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerDied","Data":"43d26e8759b6b82a7a278f2e4192f712f66101fb3e2b1a8809878a734135c75b"} Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.510206 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.686957 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687049 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687090 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687153 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687222 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687232 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687277 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687387 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.687518 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.689031 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.689442 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.695784 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvsqb\" (UniqueName: \"kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.696221 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.696314 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.696411 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs\") pod \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\" (UID: \"14955a1e-daa9-4c2c-962f-acd6eb7079ae\") " Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.696446 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.696761 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.697043 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.697095 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.697665 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.698143 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.698215 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/14955a1e-daa9-4c2c-962f-acd6eb7079ae-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.698244 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.700538 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb" (OuterVolumeSpecName: "kube-api-access-vvsqb") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "kube-api-access-vvsqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.700918 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.708080 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.754053 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.799014 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.799069 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/14955a1e-daa9-4c2c-962f-acd6eb7079ae-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.799090 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.799108 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvsqb\" (UniqueName: \"kubernetes.io/projected/14955a1e-daa9-4c2c-962f-acd6eb7079ae-kube-api-access-vvsqb\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.799126 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:52 crc kubenswrapper[4861]: I0315 00:23:52.908808 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:23:53 crc kubenswrapper[4861]: I0315 00:23:53.002259 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:53 crc kubenswrapper[4861]: I0315 00:23:53.185675 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"14955a1e-daa9-4c2c-962f-acd6eb7079ae","Type":"ContainerDied","Data":"d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e"} Mar 15 00:23:53 crc kubenswrapper[4861]: I0315 00:23:53.185749 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d891b95e094b941b5496c46f8f117a66d459fdc0eb4adac5b01e13f0d440938e" Mar 15 00:23:53 crc kubenswrapper[4861]: I0315 00:23:53.185770 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Mar 15 00:23:55 crc kubenswrapper[4861]: I0315 00:23:55.065415 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "14955a1e-daa9-4c2c-962f-acd6eb7079ae" (UID: "14955a1e-daa9-4c2c-962f-acd6eb7079ae"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:23:55 crc kubenswrapper[4861]: I0315 00:23:55.141789 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/14955a1e-daa9-4c2c-962f-acd6eb7079ae-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.917840 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:23:56 crc kubenswrapper[4861]: E0315 00:23:56.918183 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="git-clone" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918204 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="git-clone" Mar 15 00:23:56 crc kubenswrapper[4861]: E0315 00:23:56.918231 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="manage-dockerfile" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918251 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="manage-dockerfile" Mar 15 00:23:56 crc kubenswrapper[4861]: E0315 00:23:56.918274 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94246c7e-a444-40fd-b504-f71b327d289f" containerName="manage-dockerfile" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918284 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="94246c7e-a444-40fd-b504-f71b327d289f" containerName="manage-dockerfile" Mar 15 00:23:56 crc kubenswrapper[4861]: E0315 00:23:56.918298 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="docker-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918307 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="docker-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918448 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="94246c7e-a444-40fd-b504-f71b327d289f" containerName="manage-dockerfile" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.918468 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="14955a1e-daa9-4c2c-962f-acd6eb7079ae" containerName="docker-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.919346 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.922001 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.922001 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.922849 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.924026 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.931275 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970025 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbwkh\" (UniqueName: \"kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970512 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970597 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970624 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970656 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970733 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970752 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970782 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970807 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970824 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970847 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:56 crc kubenswrapper[4861]: I0315 00:23:56.970884 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.072858 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.072925 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.072962 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073005 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073042 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073077 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073135 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073184 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbwkh\" (UniqueName: \"kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073220 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073302 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073348 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073397 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073449 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073536 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073697 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.073984 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.074177 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.074250 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.074532 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.074533 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.075110 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.080671 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.082212 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.094145 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbwkh\" (UniqueName: \"kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh\") pod \"smart-gateway-operator-1-build\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.247198 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:23:57 crc kubenswrapper[4861]: I0315 00:23:57.577324 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:23:58 crc kubenswrapper[4861]: I0315 00:23:58.234448 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"7f720e09-10c4-48b0-95c8-0450137624b8","Type":"ContainerStarted","Data":"91423d745b5bd0d9129794ded43427ce9635be3ab92083d4f3bd24e8dec0ba88"} Mar 15 00:23:59 crc kubenswrapper[4861]: I0315 00:23:59.252711 4861 generic.go:334] "Generic (PLEG): container finished" podID="7f720e09-10c4-48b0-95c8-0450137624b8" containerID="c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599" exitCode=0 Mar 15 00:23:59 crc kubenswrapper[4861]: I0315 00:23:59.252789 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"7f720e09-10c4-48b0-95c8-0450137624b8","Type":"ContainerDied","Data":"c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599"} Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.150216 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558904-b8spq"] Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.151687 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.154679 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.155664 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.155724 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.158139 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558904-b8spq"] Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.218854 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7slnp\" (UniqueName: \"kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp\") pod \"auto-csr-approver-29558904-b8spq\" (UID: \"8d6e6694-d808-442b-b2af-e3d059b3a779\") " pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.263633 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"7f720e09-10c4-48b0-95c8-0450137624b8","Type":"ContainerStarted","Data":"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8"} Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.295247 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=4.295224464 podStartE2EDuration="4.295224464s" podCreationTimestamp="2026-03-15 00:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:24:00.289544438 +0000 UTC m=+1033.601741981" watchObservedRunningTime="2026-03-15 00:24:00.295224464 +0000 UTC m=+1033.607421997" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.320888 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7slnp\" (UniqueName: \"kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp\") pod \"auto-csr-approver-29558904-b8spq\" (UID: \"8d6e6694-d808-442b-b2af-e3d059b3a779\") " pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.348493 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7slnp\" (UniqueName: \"kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp\") pod \"auto-csr-approver-29558904-b8spq\" (UID: \"8d6e6694-d808-442b-b2af-e3d059b3a779\") " pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.466578 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:00 crc kubenswrapper[4861]: I0315 00:24:00.692800 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558904-b8spq"] Mar 15 00:24:01 crc kubenswrapper[4861]: I0315 00:24:01.278487 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558904-b8spq" event={"ID":"8d6e6694-d808-442b-b2af-e3d059b3a779","Type":"ContainerStarted","Data":"0bb996bd3e6b79fa141ab3add75e20302a213de174b77f840c3be7d5c2d0f040"} Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.282234 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.283326 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.285694 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.286714 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.286773 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0" gracePeriod=600 Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.288860 4861 generic.go:334] "Generic (PLEG): container finished" podID="8d6e6694-d808-442b-b2af-e3d059b3a779" containerID="4e65d81d583359399aac371fdf4d6abe208f4e396a277b6b3e769b9dabdf7874" exitCode=0 Mar 15 00:24:02 crc kubenswrapper[4861]: I0315 00:24:02.288900 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558904-b8spq" event={"ID":"8d6e6694-d808-442b-b2af-e3d059b3a779","Type":"ContainerDied","Data":"4e65d81d583359399aac371fdf4d6abe208f4e396a277b6b3e769b9dabdf7874"} Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.310136 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0" exitCode=0 Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.310213 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0"} Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.311029 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267"} Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.311060 4861 scope.go:117] "RemoveContainer" containerID="ec65e40ac95a56bce095aeec634164d385cc74f0c663da0e507ff158277e94b0" Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.595019 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.771966 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7slnp\" (UniqueName: \"kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp\") pod \"8d6e6694-d808-442b-b2af-e3d059b3a779\" (UID: \"8d6e6694-d808-442b-b2af-e3d059b3a779\") " Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.784067 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp" (OuterVolumeSpecName: "kube-api-access-7slnp") pod "8d6e6694-d808-442b-b2af-e3d059b3a779" (UID: "8d6e6694-d808-442b-b2af-e3d059b3a779"). InnerVolumeSpecName "kube-api-access-7slnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:24:03 crc kubenswrapper[4861]: I0315 00:24:03.873848 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7slnp\" (UniqueName: \"kubernetes.io/projected/8d6e6694-d808-442b-b2af-e3d059b3a779-kube-api-access-7slnp\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:04 crc kubenswrapper[4861]: I0315 00:24:04.327550 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558904-b8spq" event={"ID":"8d6e6694-d808-442b-b2af-e3d059b3a779","Type":"ContainerDied","Data":"0bb996bd3e6b79fa141ab3add75e20302a213de174b77f840c3be7d5c2d0f040"} Mar 15 00:24:04 crc kubenswrapper[4861]: I0315 00:24:04.327939 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bb996bd3e6b79fa141ab3add75e20302a213de174b77f840c3be7d5c2d0f040" Mar 15 00:24:04 crc kubenswrapper[4861]: I0315 00:24:04.327810 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558904-b8spq" Mar 15 00:24:04 crc kubenswrapper[4861]: I0315 00:24:04.664702 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558898-ptwpx"] Mar 15 00:24:04 crc kubenswrapper[4861]: I0315 00:24:04.665323 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558898-ptwpx"] Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.419142 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7ede26-54ec-44c7-ad4d-81e7bccb0251" path="/var/lib/kubelet/pods/6b7ede26-54ec-44c7-ad4d-81e7bccb0251/volumes" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.781048 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:05 crc kubenswrapper[4861]: E0315 00:24:05.781512 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6e6694-d808-442b-b2af-e3d059b3a779" containerName="oc" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.781546 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6e6694-d808-442b-b2af-e3d059b3a779" containerName="oc" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.781813 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6e6694-d808-442b-b2af-e3d059b3a779" containerName="oc" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.783420 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.800261 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.907541 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.907834 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-274vt\" (UniqueName: \"kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:05 crc kubenswrapper[4861]: I0315 00:24:05.907918 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.009134 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.009228 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-274vt\" (UniqueName: \"kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.009260 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.009750 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.009748 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.038865 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-274vt\" (UniqueName: \"kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt\") pod \"redhat-operators-xcjht\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.103244 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:06 crc kubenswrapper[4861]: I0315 00:24:06.589936 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:06 crc kubenswrapper[4861]: W0315 00:24:06.610100 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde6d93fe_e47b_4731_b928_f4b1bc05cea7.slice/crio-ca402c2568cb026751ad4ad41a39c14bf94094de207d4362cf4dd95be39289e5 WatchSource:0}: Error finding container ca402c2568cb026751ad4ad41a39c14bf94094de207d4362cf4dd95be39289e5: Status 404 returned error can't find the container with id ca402c2568cb026751ad4ad41a39c14bf94094de207d4362cf4dd95be39289e5 Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.348306 4861 generic.go:334] "Generic (PLEG): container finished" podID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerID="3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4" exitCode=0 Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.348361 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerDied","Data":"3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4"} Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.348671 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerStarted","Data":"ca402c2568cb026751ad4ad41a39c14bf94094de207d4362cf4dd95be39289e5"} Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.429146 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.429913 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="docker-build" containerID="cri-o://22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8" gracePeriod=30 Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.800538 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_7f720e09-10c4-48b0-95c8-0450137624b8/docker-build/0.log" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.801060 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941361 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941467 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbwkh\" (UniqueName: \"kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941519 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941594 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941674 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941709 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941752 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941803 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941798 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941865 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941894 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941932 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.941983 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles\") pod \"7f720e09-10c4-48b0-95c8-0450137624b8\" (UID: \"7f720e09-10c4-48b0-95c8-0450137624b8\") " Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.942328 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.942415 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.942790 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.943028 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.944317 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.944884 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.945177 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.949540 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh" (OuterVolumeSpecName: "kube-api-access-nbwkh") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "kube-api-access-nbwkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.949638 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:24:07 crc kubenswrapper[4861]: I0315 00:24:07.950939 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044791 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044854 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044871 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044895 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044910 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044922 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7f720e09-10c4-48b0-95c8-0450137624b8-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044933 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7f720e09-10c4-48b0-95c8-0450137624b8-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044946 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/7f720e09-10c4-48b0-95c8-0450137624b8-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.044957 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbwkh\" (UniqueName: \"kubernetes.io/projected/7f720e09-10c4-48b0-95c8-0450137624b8-kube-api-access-nbwkh\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.143295 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.147263 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.361638 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_7f720e09-10c4-48b0-95c8-0450137624b8/docker-build/0.log" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.362604 4861 generic.go:334] "Generic (PLEG): container finished" podID="7f720e09-10c4-48b0-95c8-0450137624b8" containerID="22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8" exitCode=1 Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.362667 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"7f720e09-10c4-48b0-95c8-0450137624b8","Type":"ContainerDied","Data":"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8"} Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.362709 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"7f720e09-10c4-48b0-95c8-0450137624b8","Type":"ContainerDied","Data":"91423d745b5bd0d9129794ded43427ce9635be3ab92083d4f3bd24e8dec0ba88"} Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.362734 4861 scope.go:117] "RemoveContainer" containerID="22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.362956 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.392368 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "7f720e09-10c4-48b0-95c8-0450137624b8" (UID: "7f720e09-10c4-48b0-95c8-0450137624b8"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.452628 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7f720e09-10c4-48b0-95c8-0450137624b8-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.462982 4861 scope.go:117] "RemoveContainer" containerID="c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.491001 4861 scope.go:117] "RemoveContainer" containerID="22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8" Mar 15 00:24:08 crc kubenswrapper[4861]: E0315 00:24:08.491669 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8\": container with ID starting with 22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8 not found: ID does not exist" containerID="22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.491738 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8"} err="failed to get container status \"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8\": rpc error: code = NotFound desc = could not find container \"22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8\": container with ID starting with 22cf4df6d79fe719e95b4fe9453d3edea8bd768da7c98289101ba8861341ffa8 not found: ID does not exist" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.491784 4861 scope.go:117] "RemoveContainer" containerID="c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599" Mar 15 00:24:08 crc kubenswrapper[4861]: E0315 00:24:08.492490 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599\": container with ID starting with c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599 not found: ID does not exist" containerID="c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.492533 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599"} err="failed to get container status \"c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599\": rpc error: code = NotFound desc = could not find container \"c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599\": container with ID starting with c2c8a5ec4d343aa67af133ac5b275429e5691c65a6cec4a14bed424d48cc7599 not found: ID does not exist" Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.708083 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:24:08 crc kubenswrapper[4861]: I0315 00:24:08.720124 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.147442 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 15 00:24:09 crc kubenswrapper[4861]: E0315 00:24:09.147881 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="docker-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.147907 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="docker-build" Mar 15 00:24:09 crc kubenswrapper[4861]: E0315 00:24:09.147935 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="manage-dockerfile" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.147949 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="manage-dockerfile" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.148145 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" containerName="docker-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.149646 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.152314 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.152379 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.153963 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.156338 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.164966 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165027 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165220 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165295 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165426 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165499 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165542 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165717 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165756 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165807 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr7c8\" (UniqueName: \"kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.165925 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.166009 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.188866 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266670 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266759 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266836 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266881 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266924 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.266968 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267007 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267083 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267085 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267122 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267270 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr7c8\" (UniqueName: \"kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267316 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267366 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.267771 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268000 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268226 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268359 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268420 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268591 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268664 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.268712 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.272545 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.274187 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.289926 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr7c8\" (UniqueName: \"kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8\") pod \"smart-gateway-operator-2-build\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.376091 4861 generic.go:334] "Generic (PLEG): container finished" podID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerID="6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753" exitCode=0 Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.376160 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerDied","Data":"6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753"} Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.420624 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f720e09-10c4-48b0-95c8-0450137624b8" path="/var/lib/kubelet/pods/7f720e09-10c4-48b0-95c8-0450137624b8/volumes" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.465747 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:24:09 crc kubenswrapper[4861]: I0315 00:24:09.787065 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Mar 15 00:24:10 crc kubenswrapper[4861]: I0315 00:24:10.383274 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerStarted","Data":"dff6b18386ecf5913157c0296fd74c07580f13a6069ffe334ec4f4074be73baa"} Mar 15 00:24:10 crc kubenswrapper[4861]: I0315 00:24:10.383687 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerStarted","Data":"2828967adfa73486d9c67c535e3f4e27786fe28337f90130da3d0699690c7f28"} Mar 15 00:24:10 crc kubenswrapper[4861]: I0315 00:24:10.386154 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerStarted","Data":"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb"} Mar 15 00:24:10 crc kubenswrapper[4861]: I0315 00:24:10.435951 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xcjht" podStartSLOduration=2.988978337 podStartE2EDuration="5.435925511s" podCreationTimestamp="2026-03-15 00:24:05 +0000 UTC" firstStartedPulling="2026-03-15 00:24:07.350344259 +0000 UTC m=+1040.662541792" lastFinishedPulling="2026-03-15 00:24:09.797291403 +0000 UTC m=+1043.109488966" observedRunningTime="2026-03-15 00:24:10.431450939 +0000 UTC m=+1043.743648482" watchObservedRunningTime="2026-03-15 00:24:10.435925511 +0000 UTC m=+1043.748123044" Mar 15 00:24:11 crc kubenswrapper[4861]: I0315 00:24:11.393226 4861 generic.go:334] "Generic (PLEG): container finished" podID="e23df747-d782-4695-88b8-382d293ceac1" containerID="dff6b18386ecf5913157c0296fd74c07580f13a6069ffe334ec4f4074be73baa" exitCode=0 Mar 15 00:24:11 crc kubenswrapper[4861]: I0315 00:24:11.393325 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerDied","Data":"dff6b18386ecf5913157c0296fd74c07580f13a6069ffe334ec4f4074be73baa"} Mar 15 00:24:12 crc kubenswrapper[4861]: I0315 00:24:12.405239 4861 generic.go:334] "Generic (PLEG): container finished" podID="e23df747-d782-4695-88b8-382d293ceac1" containerID="744acea77ec67ea91d0885b0f3cbeeb9041f48223bb594b53a8c47757fcaf881" exitCode=0 Mar 15 00:24:12 crc kubenswrapper[4861]: I0315 00:24:12.405291 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerDied","Data":"744acea77ec67ea91d0885b0f3cbeeb9041f48223bb594b53a8c47757fcaf881"} Mar 15 00:24:12 crc kubenswrapper[4861]: I0315 00:24:12.464399 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_e23df747-d782-4695-88b8-382d293ceac1/manage-dockerfile/0.log" Mar 15 00:24:13 crc kubenswrapper[4861]: I0315 00:24:13.450984 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerStarted","Data":"aa82f1fa340409b7c78978a4594a2801781b7dc3ad59b5c8afc6ad2e71a689a1"} Mar 15 00:24:13 crc kubenswrapper[4861]: I0315 00:24:13.492864 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=4.492835507 podStartE2EDuration="4.492835507s" podCreationTimestamp="2026-03-15 00:24:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:24:13.4834984 +0000 UTC m=+1046.795695963" watchObservedRunningTime="2026-03-15 00:24:13.492835507 +0000 UTC m=+1046.805033070" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.729850 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.732084 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.752147 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.869960 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.870031 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.870115 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ccwd\" (UniqueName: \"kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.971607 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.971688 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.971945 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ccwd\" (UniqueName: \"kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.972387 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.972465 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:14 crc kubenswrapper[4861]: I0315 00:24:14.997708 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ccwd\" (UniqueName: \"kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd\") pod \"certified-operators-rsdfk\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:15 crc kubenswrapper[4861]: I0315 00:24:15.062688 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:15 crc kubenswrapper[4861]: I0315 00:24:15.351458 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:15 crc kubenswrapper[4861]: I0315 00:24:15.466293 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerStarted","Data":"51e1bfa3045cf402e916b153a5d9428a0cc205907c1197d52a7cb79131f24907"} Mar 15 00:24:16 crc kubenswrapper[4861]: I0315 00:24:16.103612 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:16 crc kubenswrapper[4861]: I0315 00:24:16.103700 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:16 crc kubenswrapper[4861]: I0315 00:24:16.477192 4861 generic.go:334] "Generic (PLEG): container finished" podID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerID="6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05" exitCode=0 Mar 15 00:24:16 crc kubenswrapper[4861]: I0315 00:24:16.477245 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerDied","Data":"6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05"} Mar 15 00:24:17 crc kubenswrapper[4861]: I0315 00:24:17.158831 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xcjht" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="registry-server" probeResult="failure" output=< Mar 15 00:24:17 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:24:17 crc kubenswrapper[4861]: > Mar 15 00:24:17 crc kubenswrapper[4861]: I0315 00:24:17.486180 4861 generic.go:334] "Generic (PLEG): container finished" podID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerID="38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996" exitCode=0 Mar 15 00:24:17 crc kubenswrapper[4861]: I0315 00:24:17.486254 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerDied","Data":"38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996"} Mar 15 00:24:18 crc kubenswrapper[4861]: I0315 00:24:18.496722 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerStarted","Data":"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff"} Mar 15 00:24:18 crc kubenswrapper[4861]: I0315 00:24:18.517518 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rsdfk" podStartSLOduration=3.101237072 podStartE2EDuration="4.517495693s" podCreationTimestamp="2026-03-15 00:24:14 +0000 UTC" firstStartedPulling="2026-03-15 00:24:16.479036426 +0000 UTC m=+1049.791233959" lastFinishedPulling="2026-03-15 00:24:17.895295047 +0000 UTC m=+1051.207492580" observedRunningTime="2026-03-15 00:24:18.513551344 +0000 UTC m=+1051.825748897" watchObservedRunningTime="2026-03-15 00:24:18.517495693 +0000 UTC m=+1051.829693236" Mar 15 00:24:25 crc kubenswrapper[4861]: I0315 00:24:25.063498 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:25 crc kubenswrapper[4861]: I0315 00:24:25.064689 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:25 crc kubenswrapper[4861]: I0315 00:24:25.135819 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:25 crc kubenswrapper[4861]: I0315 00:24:25.640862 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:25 crc kubenswrapper[4861]: I0315 00:24:25.692868 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:26 crc kubenswrapper[4861]: I0315 00:24:26.162611 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:26 crc kubenswrapper[4861]: I0315 00:24:26.256985 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:27 crc kubenswrapper[4861]: I0315 00:24:27.586146 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rsdfk" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="registry-server" containerID="cri-o://0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff" gracePeriod=2 Mar 15 00:24:27 crc kubenswrapper[4861]: I0315 00:24:27.779669 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:27 crc kubenswrapper[4861]: I0315 00:24:27.779916 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xcjht" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="registry-server" containerID="cri-o://df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb" gracePeriod=2 Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.129489 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.281655 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities\") pod \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.281730 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ccwd\" (UniqueName: \"kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd\") pod \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.281767 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content\") pod \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\" (UID: \"f6ab615c-fee0-4fe5-afc9-d107b31abb46\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.283924 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities" (OuterVolumeSpecName: "utilities") pod "f6ab615c-fee0-4fe5-afc9-d107b31abb46" (UID: "f6ab615c-fee0-4fe5-afc9-d107b31abb46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.293831 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd" (OuterVolumeSpecName: "kube-api-access-7ccwd") pod "f6ab615c-fee0-4fe5-afc9-d107b31abb46" (UID: "f6ab615c-fee0-4fe5-afc9-d107b31abb46"). InnerVolumeSpecName "kube-api-access-7ccwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.318448 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.346203 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6ab615c-fee0-4fe5-afc9-d107b31abb46" (UID: "f6ab615c-fee0-4fe5-afc9-d107b31abb46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.383471 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ccwd\" (UniqueName: \"kubernetes.io/projected/f6ab615c-fee0-4fe5-afc9-d107b31abb46-kube-api-access-7ccwd\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.383509 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.383523 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6ab615c-fee0-4fe5-afc9-d107b31abb46-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.485077 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content\") pod \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.485142 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities\") pod \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.485312 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-274vt\" (UniqueName: \"kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt\") pod \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\" (UID: \"de6d93fe-e47b-4731-b928-f4b1bc05cea7\") " Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.487727 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities" (OuterVolumeSpecName: "utilities") pod "de6d93fe-e47b-4731-b928-f4b1bc05cea7" (UID: "de6d93fe-e47b-4731-b928-f4b1bc05cea7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.489653 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt" (OuterVolumeSpecName: "kube-api-access-274vt") pod "de6d93fe-e47b-4731-b928-f4b1bc05cea7" (UID: "de6d93fe-e47b-4731-b928-f4b1bc05cea7"). InnerVolumeSpecName "kube-api-access-274vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.586917 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-274vt\" (UniqueName: \"kubernetes.io/projected/de6d93fe-e47b-4731-b928-f4b1bc05cea7-kube-api-access-274vt\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.586965 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.594939 4861 generic.go:334] "Generic (PLEG): container finished" podID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerID="0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff" exitCode=0 Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.595064 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerDied","Data":"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff"} Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.595103 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rsdfk" event={"ID":"f6ab615c-fee0-4fe5-afc9-d107b31abb46","Type":"ContainerDied","Data":"51e1bfa3045cf402e916b153a5d9428a0cc205907c1197d52a7cb79131f24907"} Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.595128 4861 scope.go:117] "RemoveContainer" containerID="0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.595302 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rsdfk" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.600684 4861 generic.go:334] "Generic (PLEG): container finished" podID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerID="df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb" exitCode=0 Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.600743 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerDied","Data":"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb"} Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.600783 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xcjht" event={"ID":"de6d93fe-e47b-4731-b928-f4b1bc05cea7","Type":"ContainerDied","Data":"ca402c2568cb026751ad4ad41a39c14bf94094de207d4362cf4dd95be39289e5"} Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.600873 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xcjht" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.617365 4861 scope.go:117] "RemoveContainer" containerID="38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.629953 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de6d93fe-e47b-4731-b928-f4b1bc05cea7" (UID: "de6d93fe-e47b-4731-b928-f4b1bc05cea7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.636444 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.641531 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rsdfk"] Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.652986 4861 scope.go:117] "RemoveContainer" containerID="6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.677156 4861 scope.go:117] "RemoveContainer" containerID="0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.677740 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff\": container with ID starting with 0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff not found: ID does not exist" containerID="0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.677770 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff"} err="failed to get container status \"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff\": rpc error: code = NotFound desc = could not find container \"0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff\": container with ID starting with 0897a39d87428b6754d801da959abbb01b4cb19116d0618d03e557b866f507ff not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.677792 4861 scope.go:117] "RemoveContainer" containerID="38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.678125 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996\": container with ID starting with 38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996 not found: ID does not exist" containerID="38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.678157 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996"} err="failed to get container status \"38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996\": rpc error: code = NotFound desc = could not find container \"38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996\": container with ID starting with 38cdaadce30c3cb726c99d964ca1c4e91ecaed23be991bbc8da0db1ca986d996 not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.678172 4861 scope.go:117] "RemoveContainer" containerID="6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.678673 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05\": container with ID starting with 6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05 not found: ID does not exist" containerID="6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.678692 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05"} err="failed to get container status \"6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05\": rpc error: code = NotFound desc = could not find container \"6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05\": container with ID starting with 6f11aca580eeba832675323842e87478e63602dd1372451266982f2148dc6b05 not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.678705 4861 scope.go:117] "RemoveContainer" containerID="df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.687892 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de6d93fe-e47b-4731-b928-f4b1bc05cea7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.695923 4861 scope.go:117] "RemoveContainer" containerID="6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.712518 4861 scope.go:117] "RemoveContainer" containerID="3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.740761 4861 scope.go:117] "RemoveContainer" containerID="df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.742115 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb\": container with ID starting with df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb not found: ID does not exist" containerID="df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.742145 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb"} err="failed to get container status \"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb\": rpc error: code = NotFound desc = could not find container \"df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb\": container with ID starting with df2be4c9127548b749d288a0bc72776340ce2cbdf7c59faad5bf03d3c024affb not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.742166 4861 scope.go:117] "RemoveContainer" containerID="6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.742611 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753\": container with ID starting with 6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753 not found: ID does not exist" containerID="6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.742634 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753"} err="failed to get container status \"6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753\": rpc error: code = NotFound desc = could not find container \"6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753\": container with ID starting with 6f6f94775af714e5ffc6637d5ce4df370a5a19b0267b61dd0ec42c94a6e96753 not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.742650 4861 scope.go:117] "RemoveContainer" containerID="3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4" Mar 15 00:24:28 crc kubenswrapper[4861]: E0315 00:24:28.744092 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4\": container with ID starting with 3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4 not found: ID does not exist" containerID="3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.744120 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4"} err="failed to get container status \"3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4\": rpc error: code = NotFound desc = could not find container \"3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4\": container with ID starting with 3186dbd4360bf6f375bcb350da9380f0f9eb20397af9bbeebb3599d3eae3a7c4 not found: ID does not exist" Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.951595 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:28 crc kubenswrapper[4861]: I0315 00:24:28.956625 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xcjht"] Mar 15 00:24:29 crc kubenswrapper[4861]: I0315 00:24:29.435159 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" path="/var/lib/kubelet/pods/de6d93fe-e47b-4731-b928-f4b1bc05cea7/volumes" Mar 15 00:24:29 crc kubenswrapper[4861]: I0315 00:24:29.436419 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" path="/var/lib/kubelet/pods/f6ab615c-fee0-4fe5-afc9-d107b31abb46/volumes" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.897063 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898142 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="extract-content" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898159 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="extract-content" Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898178 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="extract-utilities" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898186 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="extract-utilities" Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898197 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="extract-utilities" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898204 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="extract-utilities" Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898227 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898234 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898249 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="extract-content" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898256 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="extract-content" Mar 15 00:24:33 crc kubenswrapper[4861]: E0315 00:24:33.898273 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898280 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898424 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6d93fe-e47b-4731-b928-f4b1bc05cea7" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.898433 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ab615c-fee0-4fe5-afc9-d107b31abb46" containerName="registry-server" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.899460 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:33 crc kubenswrapper[4861]: I0315 00:24:33.908405 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.065479 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.065545 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgckc\" (UniqueName: \"kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.065826 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.167221 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.167317 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.167362 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgckc\" (UniqueName: \"kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.167838 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.167921 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.191706 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgckc\" (UniqueName: \"kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc\") pod \"community-operators-l59gf\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.282698 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.554207 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:34 crc kubenswrapper[4861]: W0315 00:24:34.564132 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd9d58d6_76ce_477c_9f2d_a0444838033d.slice/crio-a23ca4c6f61c704211b92d035ea665ae24f898aa335612faab9efbd7a515339a WatchSource:0}: Error finding container a23ca4c6f61c704211b92d035ea665ae24f898aa335612faab9efbd7a515339a: Status 404 returned error can't find the container with id a23ca4c6f61c704211b92d035ea665ae24f898aa335612faab9efbd7a515339a Mar 15 00:24:34 crc kubenswrapper[4861]: I0315 00:24:34.650794 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerStarted","Data":"a23ca4c6f61c704211b92d035ea665ae24f898aa335612faab9efbd7a515339a"} Mar 15 00:24:35 crc kubenswrapper[4861]: I0315 00:24:35.659472 4861 generic.go:334] "Generic (PLEG): container finished" podID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerID="187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b" exitCode=0 Mar 15 00:24:35 crc kubenswrapper[4861]: I0315 00:24:35.659593 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerDied","Data":"187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b"} Mar 15 00:24:37 crc kubenswrapper[4861]: I0315 00:24:37.683446 4861 generic.go:334] "Generic (PLEG): container finished" podID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerID="4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b" exitCode=0 Mar 15 00:24:37 crc kubenswrapper[4861]: I0315 00:24:37.683826 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerDied","Data":"4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b"} Mar 15 00:24:38 crc kubenswrapper[4861]: I0315 00:24:38.692971 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerStarted","Data":"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1"} Mar 15 00:24:38 crc kubenswrapper[4861]: I0315 00:24:38.716624 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l59gf" podStartSLOduration=3.30823377 podStartE2EDuration="5.716598393s" podCreationTimestamp="2026-03-15 00:24:33 +0000 UTC" firstStartedPulling="2026-03-15 00:24:35.661371575 +0000 UTC m=+1068.973569108" lastFinishedPulling="2026-03-15 00:24:38.069736198 +0000 UTC m=+1071.381933731" observedRunningTime="2026-03-15 00:24:38.710716412 +0000 UTC m=+1072.022913955" watchObservedRunningTime="2026-03-15 00:24:38.716598393 +0000 UTC m=+1072.028795926" Mar 15 00:24:44 crc kubenswrapper[4861]: I0315 00:24:44.283513 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:44 crc kubenswrapper[4861]: I0315 00:24:44.284037 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:44 crc kubenswrapper[4861]: I0315 00:24:44.336250 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:44 crc kubenswrapper[4861]: I0315 00:24:44.779373 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:45 crc kubenswrapper[4861]: I0315 00:24:45.581646 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:46 crc kubenswrapper[4861]: I0315 00:24:46.754187 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l59gf" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="registry-server" containerID="cri-o://6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1" gracePeriod=2 Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.761892 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.764669 4861 generic.go:334] "Generic (PLEG): container finished" podID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerID="6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1" exitCode=0 Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.764717 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerDied","Data":"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1"} Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.764746 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l59gf" event={"ID":"bd9d58d6-76ce-477c-9f2d-a0444838033d","Type":"ContainerDied","Data":"a23ca4c6f61c704211b92d035ea665ae24f898aa335612faab9efbd7a515339a"} Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.764767 4861 scope.go:117] "RemoveContainer" containerID="6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.804751 4861 scope.go:117] "RemoveContainer" containerID="4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.830481 4861 scope.go:117] "RemoveContainer" containerID="187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.850166 4861 scope.go:117] "RemoveContainer" containerID="6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1" Mar 15 00:24:47 crc kubenswrapper[4861]: E0315 00:24:47.852173 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1\": container with ID starting with 6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1 not found: ID does not exist" containerID="6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.852353 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1"} err="failed to get container status \"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1\": rpc error: code = NotFound desc = could not find container \"6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1\": container with ID starting with 6609658233c602ad64f4a4096c28d6f810b38ded8558d1b82ee1e6d887f592f1 not found: ID does not exist" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.852506 4861 scope.go:117] "RemoveContainer" containerID="4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b" Mar 15 00:24:47 crc kubenswrapper[4861]: E0315 00:24:47.853016 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b\": container with ID starting with 4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b not found: ID does not exist" containerID="4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.853178 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b"} err="failed to get container status \"4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b\": rpc error: code = NotFound desc = could not find container \"4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b\": container with ID starting with 4f7106a170629d8750490b261d168fefe626dc0989aa7ca396814dcc4117326b not found: ID does not exist" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.853346 4861 scope.go:117] "RemoveContainer" containerID="187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b" Mar 15 00:24:47 crc kubenswrapper[4861]: E0315 00:24:47.854074 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b\": container with ID starting with 187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b not found: ID does not exist" containerID="187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.854154 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b"} err="failed to get container status \"187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b\": rpc error: code = NotFound desc = could not find container \"187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b\": container with ID starting with 187bfc5368841c779d07481a6463fa45ba47fbaa3127e81faf220cb9fa196e1b not found: ID does not exist" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.874419 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgckc\" (UniqueName: \"kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc\") pod \"bd9d58d6-76ce-477c-9f2d-a0444838033d\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.874477 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content\") pod \"bd9d58d6-76ce-477c-9f2d-a0444838033d\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.875018 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities\") pod \"bd9d58d6-76ce-477c-9f2d-a0444838033d\" (UID: \"bd9d58d6-76ce-477c-9f2d-a0444838033d\") " Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.877835 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities" (OuterVolumeSpecName: "utilities") pod "bd9d58d6-76ce-477c-9f2d-a0444838033d" (UID: "bd9d58d6-76ce-477c-9f2d-a0444838033d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.882967 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc" (OuterVolumeSpecName: "kube-api-access-dgckc") pod "bd9d58d6-76ce-477c-9f2d-a0444838033d" (UID: "bd9d58d6-76ce-477c-9f2d-a0444838033d"). InnerVolumeSpecName "kube-api-access-dgckc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.945040 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd9d58d6-76ce-477c-9f2d-a0444838033d" (UID: "bd9d58d6-76ce-477c-9f2d-a0444838033d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.976829 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.976875 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgckc\" (UniqueName: \"kubernetes.io/projected/bd9d58d6-76ce-477c-9f2d-a0444838033d-kube-api-access-dgckc\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:47 crc kubenswrapper[4861]: I0315 00:24:47.976908 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd9d58d6-76ce-477c-9f2d-a0444838033d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:24:48 crc kubenswrapper[4861]: I0315 00:24:48.180551 4861 scope.go:117] "RemoveContainer" containerID="5456683434de2f72fd984637411a566af681cd1a7101a6b34dcb99112b350e91" Mar 15 00:24:48 crc kubenswrapper[4861]: I0315 00:24:48.770921 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l59gf" Mar 15 00:24:48 crc kubenswrapper[4861]: I0315 00:24:48.810726 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:48 crc kubenswrapper[4861]: I0315 00:24:48.820245 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l59gf"] Mar 15 00:24:49 crc kubenswrapper[4861]: I0315 00:24:49.423877 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" path="/var/lib/kubelet/pods/bd9d58d6-76ce-477c-9f2d-a0444838033d/volumes" Mar 15 00:25:23 crc kubenswrapper[4861]: I0315 00:25:23.063347 4861 generic.go:334] "Generic (PLEG): container finished" podID="e23df747-d782-4695-88b8-382d293ceac1" containerID="aa82f1fa340409b7c78978a4594a2801781b7dc3ad59b5c8afc6ad2e71a689a1" exitCode=0 Mar 15 00:25:23 crc kubenswrapper[4861]: I0315 00:25:23.063438 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerDied","Data":"aa82f1fa340409b7c78978a4594a2801781b7dc3ad59b5c8afc6ad2e71a689a1"} Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.418661 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.586772 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.586938 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.586970 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587012 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587044 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587084 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587132 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587147 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587178 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587264 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587533 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587803 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587875 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.587922 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr7c8\" (UniqueName: \"kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8\") pod \"e23df747-d782-4695-88b8-382d293ceac1\" (UID: \"e23df747-d782-4695-88b8-382d293ceac1\") " Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588383 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588733 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588788 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588809 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e23df747-d782-4695-88b8-382d293ceac1-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588721 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.588822 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.589306 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.595996 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.597912 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.597902 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8" (OuterVolumeSpecName: "kube-api-access-hr7c8") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "kube-api-access-hr7c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.598861 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690054 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690121 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690141 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/e23df747-d782-4695-88b8-382d293ceac1-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690159 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690173 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e23df747-d782-4695-88b8-382d293ceac1-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690186 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.690198 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr7c8\" (UniqueName: \"kubernetes.io/projected/e23df747-d782-4695-88b8-382d293ceac1-kube-api-access-hr7c8\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.768420 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:24 crc kubenswrapper[4861]: I0315 00:25:24.792109 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:25 crc kubenswrapper[4861]: I0315 00:25:25.082855 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"e23df747-d782-4695-88b8-382d293ceac1","Type":"ContainerDied","Data":"2828967adfa73486d9c67c535e3f4e27786fe28337f90130da3d0699690c7f28"} Mar 15 00:25:25 crc kubenswrapper[4861]: I0315 00:25:25.082909 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2828967adfa73486d9c67c535e3f4e27786fe28337f90130da3d0699690c7f28" Mar 15 00:25:25 crc kubenswrapper[4861]: I0315 00:25:25.082982 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Mar 15 00:25:26 crc kubenswrapper[4861]: I0315 00:25:26.672052 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e23df747-d782-4695-88b8-382d293ceac1" (UID: "e23df747-d782-4695-88b8-382d293ceac1"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:26 crc kubenswrapper[4861]: I0315 00:25:26.725238 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e23df747-d782-4695-88b8-382d293ceac1-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950362 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950779 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="registry-server" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950799 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="registry-server" Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950815 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="git-clone" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950825 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="git-clone" Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950835 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="extract-content" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950845 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="extract-content" Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950864 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="manage-dockerfile" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950873 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="manage-dockerfile" Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950884 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="docker-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950892 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="docker-build" Mar 15 00:25:28 crc kubenswrapper[4861]: E0315 00:25:28.950923 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="extract-utilities" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.950931 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="extract-utilities" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.951073 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23df747-d782-4695-88b8-382d293ceac1" containerName="docker-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.951090 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd9d58d6-76ce-477c-9f2d-a0444838033d" containerName="registry-server" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.951978 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.956074 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.956504 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.956602 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.956663 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjkgl\" (UniqueName: \"kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957106 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957146 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957159 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957378 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957478 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957520 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957784 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957835 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957878 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.957932 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.959310 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.960273 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Mar 15 00:25:28 crc kubenswrapper[4861]: I0315 00:25:28.971718 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059514 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059627 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059668 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjkgl\" (UniqueName: \"kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059728 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059758 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059857 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059895 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059930 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.059968 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.060004 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.060036 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.060080 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.060480 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.061035 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.061195 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.061247 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.061424 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.061577 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.071905 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.073640 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.582343 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.583216 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.583246 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.589609 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjkgl\" (UniqueName: \"kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl\") pod \"sg-core-1-build\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.616319 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 15 00:25:29 crc kubenswrapper[4861]: I0315 00:25:29.865703 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:30 crc kubenswrapper[4861]: I0315 00:25:30.124055 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3","Type":"ContainerStarted","Data":"542980cc5638ab93932b7db2ebcf5984941d37b2e3a6469b2f940d5c8e3a9ed8"} Mar 15 00:25:31 crc kubenswrapper[4861]: I0315 00:25:31.134954 4861 generic.go:334] "Generic (PLEG): container finished" podID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerID="2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b" exitCode=0 Mar 15 00:25:31 crc kubenswrapper[4861]: I0315 00:25:31.135031 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3","Type":"ContainerDied","Data":"2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b"} Mar 15 00:25:32 crc kubenswrapper[4861]: I0315 00:25:32.149294 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3","Type":"ContainerStarted","Data":"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29"} Mar 15 00:25:32 crc kubenswrapper[4861]: I0315 00:25:32.187342 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=4.187316021 podStartE2EDuration="4.187316021s" podCreationTimestamp="2026-03-15 00:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:25:32.177875991 +0000 UTC m=+1125.490073564" watchObservedRunningTime="2026-03-15 00:25:32.187316021 +0000 UTC m=+1125.499513584" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.309146 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.310442 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="docker-build" containerID="cri-o://6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29" gracePeriod=30 Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.691976 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_b7d3a465-7c21-4ac2-b1b1-7aab347d25e3/docker-build/0.log" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.693036 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787248 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787431 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787470 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787509 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787551 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787607 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787642 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787677 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjkgl\" (UniqueName: \"kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.787744 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.789305 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.790089 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.796483 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.796921 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl" (OuterVolumeSpecName: "kube-api-access-zjkgl") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "kube-api-access-zjkgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.796965 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.797252 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.797304 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.797334 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.797361 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.797388 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjkgl\" (UniqueName: \"kubernetes.io/projected/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-kube-api-access-zjkgl\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.798212 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.801393 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.898672 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.898765 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.898811 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir\") pod \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\" (UID: \"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3\") " Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899106 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899495 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899533 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899359 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899586 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.899925 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.917472 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:39 crc kubenswrapper[4861]: I0315 00:25:39.972687 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" (UID: "b7d3a465-7c21-4ac2-b1b1-7aab347d25e3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.000664 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.000713 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.000725 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.000736 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.225939 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_b7d3a465-7c21-4ac2-b1b1-7aab347d25e3/docker-build/0.log" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.227408 4861 generic.go:334] "Generic (PLEG): container finished" podID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerID="6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29" exitCode=1 Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.227474 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3","Type":"ContainerDied","Data":"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29"} Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.227528 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.227533 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"b7d3a465-7c21-4ac2-b1b1-7aab347d25e3","Type":"ContainerDied","Data":"542980cc5638ab93932b7db2ebcf5984941d37b2e3a6469b2f940d5c8e3a9ed8"} Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.227576 4861 scope.go:117] "RemoveContainer" containerID="6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.280735 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.291720 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.304943 4861 scope.go:117] "RemoveContainer" containerID="2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.335411 4861 scope.go:117] "RemoveContainer" containerID="6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29" Mar 15 00:25:40 crc kubenswrapper[4861]: E0315 00:25:40.336166 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29\": container with ID starting with 6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29 not found: ID does not exist" containerID="6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.336243 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29"} err="failed to get container status \"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29\": rpc error: code = NotFound desc = could not find container \"6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29\": container with ID starting with 6f965be0aba008b0466d4907b5c2ed809e39c0b677eede76c040e9ee3bd15e29 not found: ID does not exist" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.336294 4861 scope.go:117] "RemoveContainer" containerID="2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b" Mar 15 00:25:40 crc kubenswrapper[4861]: E0315 00:25:40.337018 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b\": container with ID starting with 2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b not found: ID does not exist" containerID="2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.337152 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b"} err="failed to get container status \"2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b\": rpc error: code = NotFound desc = could not find container \"2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b\": container with ID starting with 2897c5f237bc153b9c4710f64da2f3920cb87bd30422572f9643559e3de3b29b not found: ID does not exist" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.973456 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 15 00:25:40 crc kubenswrapper[4861]: E0315 00:25:40.973867 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="docker-build" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.973888 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="docker-build" Mar 15 00:25:40 crc kubenswrapper[4861]: E0315 00:25:40.973910 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="manage-dockerfile" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.973919 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="manage-dockerfile" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.974062 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" containerName="docker-build" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.975121 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.978223 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.978292 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.978783 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.978838 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Mar 15 00:25:40 crc kubenswrapper[4861]: I0315 00:25:40.999669 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012733 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012796 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012823 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012847 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbwsj\" (UniqueName: \"kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012878 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012905 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012933 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012957 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.012986 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.013014 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.013056 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.013094 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114427 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114518 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114605 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114649 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114691 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114768 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114839 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114882 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.114885 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115353 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115438 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115509 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115622 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbwsj\" (UniqueName: \"kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115635 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115661 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115532 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115703 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115710 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.115974 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.116758 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.117064 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.133048 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.133261 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.138395 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbwsj\" (UniqueName: \"kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj\") pod \"sg-core-2-build\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.309254 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.419814 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7d3a465-7c21-4ac2-b1b1-7aab347d25e3" path="/var/lib/kubelet/pods/b7d3a465-7c21-4ac2-b1b1-7aab347d25e3/volumes" Mar 15 00:25:41 crc kubenswrapper[4861]: I0315 00:25:41.637608 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Mar 15 00:25:42 crc kubenswrapper[4861]: I0315 00:25:42.252993 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerStarted","Data":"1c95bb0e6cc02a6d8a4a76bfeb07ce42c24f0d455f971f003fbc62d0d13b8c6e"} Mar 15 00:25:42 crc kubenswrapper[4861]: I0315 00:25:42.253075 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerStarted","Data":"1b903099d9c67fa5032ac2d5c9c6c80be5dc6d19d0cd6e70506e183054e578ac"} Mar 15 00:25:43 crc kubenswrapper[4861]: I0315 00:25:43.264953 4861 generic.go:334] "Generic (PLEG): container finished" podID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerID="1c95bb0e6cc02a6d8a4a76bfeb07ce42c24f0d455f971f003fbc62d0d13b8c6e" exitCode=0 Mar 15 00:25:43 crc kubenswrapper[4861]: I0315 00:25:43.265061 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerDied","Data":"1c95bb0e6cc02a6d8a4a76bfeb07ce42c24f0d455f971f003fbc62d0d13b8c6e"} Mar 15 00:25:44 crc kubenswrapper[4861]: I0315 00:25:44.274714 4861 generic.go:334] "Generic (PLEG): container finished" podID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerID="9e5f6410bc358d0b4d2c2135e0fb5fa0f33a3da1643b521bcc35386dcbf8e342" exitCode=0 Mar 15 00:25:44 crc kubenswrapper[4861]: I0315 00:25:44.274770 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerDied","Data":"9e5f6410bc358d0b4d2c2135e0fb5fa0f33a3da1643b521bcc35386dcbf8e342"} Mar 15 00:25:44 crc kubenswrapper[4861]: I0315 00:25:44.319498 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_ba3891b9-a262-4c34-93bf-fd129e5e0b1d/manage-dockerfile/0.log" Mar 15 00:25:45 crc kubenswrapper[4861]: I0315 00:25:45.289738 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerStarted","Data":"455dc19bf89adfc34c68aa4cdbe9c5181974995a13420411d8bd87377b34d9a6"} Mar 15 00:25:45 crc kubenswrapper[4861]: I0315 00:25:45.331629 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=5.33159882 podStartE2EDuration="5.33159882s" podCreationTimestamp="2026-03-15 00:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:25:45.328235409 +0000 UTC m=+1138.640432942" watchObservedRunningTime="2026-03-15 00:25:45.33159882 +0000 UTC m=+1138.643796393" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.151854 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558906-slfcz"] Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.153978 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.157491 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.157814 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.158063 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.162243 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558906-slfcz"] Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.207068 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm4ch\" (UniqueName: \"kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch\") pod \"auto-csr-approver-29558906-slfcz\" (UID: \"fdd1741e-a90b-4025-ac04-e1da8c46ca7f\") " pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.308650 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm4ch\" (UniqueName: \"kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch\") pod \"auto-csr-approver-29558906-slfcz\" (UID: \"fdd1741e-a90b-4025-ac04-e1da8c46ca7f\") " pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.329477 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm4ch\" (UniqueName: \"kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch\") pod \"auto-csr-approver-29558906-slfcz\" (UID: \"fdd1741e-a90b-4025-ac04-e1da8c46ca7f\") " pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.472249 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:00 crc kubenswrapper[4861]: I0315 00:26:00.959674 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558906-slfcz"] Mar 15 00:26:00 crc kubenswrapper[4861]: W0315 00:26:00.979612 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdd1741e_a90b_4025_ac04_e1da8c46ca7f.slice/crio-3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6 WatchSource:0}: Error finding container 3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6: Status 404 returned error can't find the container with id 3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6 Mar 15 00:26:01 crc kubenswrapper[4861]: I0315 00:26:01.426250 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558906-slfcz" event={"ID":"fdd1741e-a90b-4025-ac04-e1da8c46ca7f","Type":"ContainerStarted","Data":"3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6"} Mar 15 00:26:02 crc kubenswrapper[4861]: I0315 00:26:02.282415 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:26:02 crc kubenswrapper[4861]: I0315 00:26:02.283062 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:26:02 crc kubenswrapper[4861]: I0315 00:26:02.438606 4861 generic.go:334] "Generic (PLEG): container finished" podID="fdd1741e-a90b-4025-ac04-e1da8c46ca7f" containerID="3ec4a821e25757a09d89c7e435091981f55950e1ca0a5dbcddaee3b19deabbc3" exitCode=0 Mar 15 00:26:02 crc kubenswrapper[4861]: I0315 00:26:02.438688 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558906-slfcz" event={"ID":"fdd1741e-a90b-4025-ac04-e1da8c46ca7f","Type":"ContainerDied","Data":"3ec4a821e25757a09d89c7e435091981f55950e1ca0a5dbcddaee3b19deabbc3"} Mar 15 00:26:03 crc kubenswrapper[4861]: I0315 00:26:03.751537 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:03 crc kubenswrapper[4861]: I0315 00:26:03.775121 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm4ch\" (UniqueName: \"kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch\") pod \"fdd1741e-a90b-4025-ac04-e1da8c46ca7f\" (UID: \"fdd1741e-a90b-4025-ac04-e1da8c46ca7f\") " Mar 15 00:26:03 crc kubenswrapper[4861]: I0315 00:26:03.796499 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch" (OuterVolumeSpecName: "kube-api-access-mm4ch") pod "fdd1741e-a90b-4025-ac04-e1da8c46ca7f" (UID: "fdd1741e-a90b-4025-ac04-e1da8c46ca7f"). InnerVolumeSpecName "kube-api-access-mm4ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:26:03 crc kubenswrapper[4861]: I0315 00:26:03.877445 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm4ch\" (UniqueName: \"kubernetes.io/projected/fdd1741e-a90b-4025-ac04-e1da8c46ca7f-kube-api-access-mm4ch\") on node \"crc\" DevicePath \"\"" Mar 15 00:26:04 crc kubenswrapper[4861]: I0315 00:26:04.460074 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558906-slfcz" event={"ID":"fdd1741e-a90b-4025-ac04-e1da8c46ca7f","Type":"ContainerDied","Data":"3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6"} Mar 15 00:26:04 crc kubenswrapper[4861]: I0315 00:26:04.460598 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3894fdafea66bf213fef7eae25512960b681a3e11dc983655b97ed1efed01fc6" Mar 15 00:26:04 crc kubenswrapper[4861]: I0315 00:26:04.460161 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558906-slfcz" Mar 15 00:26:04 crc kubenswrapper[4861]: I0315 00:26:04.814538 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558900-dm476"] Mar 15 00:26:04 crc kubenswrapper[4861]: I0315 00:26:04.823124 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558900-dm476"] Mar 15 00:26:05 crc kubenswrapper[4861]: I0315 00:26:05.418509 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34e35417-c6f9-47c0-84c9-0ef2f6482447" path="/var/lib/kubelet/pods/34e35417-c6f9-47c0-84c9-0ef2f6482447/volumes" Mar 15 00:26:32 crc kubenswrapper[4861]: I0315 00:26:32.282474 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:26:32 crc kubenswrapper[4861]: I0315 00:26:32.283504 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:26:48 crc kubenswrapper[4861]: I0315 00:26:48.338426 4861 scope.go:117] "RemoveContainer" containerID="00551e8d857c4aa507b818b487aaef5d382d34ed4b821ee965749493f7db6537" Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.281616 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.282139 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.282185 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.282812 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.282860 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267" gracePeriod=600 Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.922699 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267" exitCode=0 Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.922770 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267"} Mar 15 00:27:02 crc kubenswrapper[4861]: I0315 00:27:02.922854 4861 scope.go:117] "RemoveContainer" containerID="af405c259669db354b4603f01d4c67588ce45cfd98cba530cd99ffb285a18ef0" Mar 15 00:27:03 crc kubenswrapper[4861]: I0315 00:27:03.932465 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19"} Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.151594 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558908-6l8x4"] Mar 15 00:28:00 crc kubenswrapper[4861]: E0315 00:28:00.152674 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd1741e-a90b-4025-ac04-e1da8c46ca7f" containerName="oc" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.152694 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd1741e-a90b-4025-ac04-e1da8c46ca7f" containerName="oc" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.152846 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd1741e-a90b-4025-ac04-e1da8c46ca7f" containerName="oc" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.153386 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.156601 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.156808 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.156943 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.163218 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558908-6l8x4"] Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.267806 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch8pk\" (UniqueName: \"kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk\") pod \"auto-csr-approver-29558908-6l8x4\" (UID: \"a1ef5f4c-e426-43d2-90f8-c21be80ece91\") " pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.369239 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch8pk\" (UniqueName: \"kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk\") pod \"auto-csr-approver-29558908-6l8x4\" (UID: \"a1ef5f4c-e426-43d2-90f8-c21be80ece91\") " pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.392351 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch8pk\" (UniqueName: \"kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk\") pod \"auto-csr-approver-29558908-6l8x4\" (UID: \"a1ef5f4c-e426-43d2-90f8-c21be80ece91\") " pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.483266 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.941744 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558908-6l8x4"] Mar 15 00:28:00 crc kubenswrapper[4861]: I0315 00:28:00.949640 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:28:01 crc kubenswrapper[4861]: I0315 00:28:01.370421 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" event={"ID":"a1ef5f4c-e426-43d2-90f8-c21be80ece91","Type":"ContainerStarted","Data":"3cdb3d1303136fc4be1dd9ef5861c886bd3ff8d1621f721857758b687630f1dc"} Mar 15 00:28:03 crc kubenswrapper[4861]: I0315 00:28:03.389183 4861 generic.go:334] "Generic (PLEG): container finished" podID="a1ef5f4c-e426-43d2-90f8-c21be80ece91" containerID="b5dd2c42d411ad4fc426dfd17121ea463207a4b2c8d151104992489dd91731a3" exitCode=0 Mar 15 00:28:03 crc kubenswrapper[4861]: I0315 00:28:03.389254 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" event={"ID":"a1ef5f4c-e426-43d2-90f8-c21be80ece91","Type":"ContainerDied","Data":"b5dd2c42d411ad4fc426dfd17121ea463207a4b2c8d151104992489dd91731a3"} Mar 15 00:28:04 crc kubenswrapper[4861]: I0315 00:28:04.700057 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:04 crc kubenswrapper[4861]: I0315 00:28:04.836728 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch8pk\" (UniqueName: \"kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk\") pod \"a1ef5f4c-e426-43d2-90f8-c21be80ece91\" (UID: \"a1ef5f4c-e426-43d2-90f8-c21be80ece91\") " Mar 15 00:28:04 crc kubenswrapper[4861]: I0315 00:28:04.858850 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk" (OuterVolumeSpecName: "kube-api-access-ch8pk") pod "a1ef5f4c-e426-43d2-90f8-c21be80ece91" (UID: "a1ef5f4c-e426-43d2-90f8-c21be80ece91"). InnerVolumeSpecName "kube-api-access-ch8pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:28:04 crc kubenswrapper[4861]: I0315 00:28:04.940338 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch8pk\" (UniqueName: \"kubernetes.io/projected/a1ef5f4c-e426-43d2-90f8-c21be80ece91-kube-api-access-ch8pk\") on node \"crc\" DevicePath \"\"" Mar 15 00:28:05 crc kubenswrapper[4861]: I0315 00:28:05.405694 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" event={"ID":"a1ef5f4c-e426-43d2-90f8-c21be80ece91","Type":"ContainerDied","Data":"3cdb3d1303136fc4be1dd9ef5861c886bd3ff8d1621f721857758b687630f1dc"} Mar 15 00:28:05 crc kubenswrapper[4861]: I0315 00:28:05.405771 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cdb3d1303136fc4be1dd9ef5861c886bd3ff8d1621f721857758b687630f1dc" Mar 15 00:28:05 crc kubenswrapper[4861]: I0315 00:28:05.405836 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558908-6l8x4" Mar 15 00:28:05 crc kubenswrapper[4861]: I0315 00:28:05.785768 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558902-46bp7"] Mar 15 00:28:05 crc kubenswrapper[4861]: I0315 00:28:05.794970 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558902-46bp7"] Mar 15 00:28:07 crc kubenswrapper[4861]: I0315 00:28:07.418057 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b38eb7a-7876-433d-b110-7030850e9252" path="/var/lib/kubelet/pods/5b38eb7a-7876-433d-b110-7030850e9252/volumes" Mar 15 00:28:48 crc kubenswrapper[4861]: I0315 00:28:48.457768 4861 scope.go:117] "RemoveContainer" containerID="0f5eed031a8d9e493499f3b34294fbd558f33718d3293b45ca5e72747cc503a2" Mar 15 00:29:04 crc kubenswrapper[4861]: I0315 00:29:04.897854 4861 generic.go:334] "Generic (PLEG): container finished" podID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerID="455dc19bf89adfc34c68aa4cdbe9c5181974995a13420411d8bd87377b34d9a6" exitCode=0 Mar 15 00:29:04 crc kubenswrapper[4861]: I0315 00:29:04.897985 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerDied","Data":"455dc19bf89adfc34c68aa4cdbe9c5181974995a13420411d8bd87377b34d9a6"} Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.217030 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.274835 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.274910 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbwsj\" (UniqueName: \"kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.274948 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275005 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275034 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275055 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275085 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275112 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275141 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275158 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275244 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275280 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root\") pod \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\" (UID: \"ba3891b9-a262-4c34-93bf-fd129e5e0b1d\") " Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275453 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275750 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275844 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.275969 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.276472 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.276692 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.277007 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.291226 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.292638 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj" (OuterVolumeSpecName: "kube-api-access-hbwsj") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "kube-api-access-hbwsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.292694 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.292746 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.377765 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378295 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378308 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378320 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378332 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378342 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbwsj\" (UniqueName: \"kubernetes.io/projected/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-kube-api-access-hbwsj\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378352 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378364 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.378375 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.927657 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"ba3891b9-a262-4c34-93bf-fd129e5e0b1d","Type":"ContainerDied","Data":"1b903099d9c67fa5032ac2d5c9c6c80be5dc6d19d0cd6e70506e183054e578ac"} Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.928102 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b903099d9c67fa5032ac2d5c9c6c80be5dc6d19d0cd6e70506e183054e578ac" Mar 15 00:29:06 crc kubenswrapper[4861]: I0315 00:29:06.927818 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Mar 15 00:29:07 crc kubenswrapper[4861]: I0315 00:29:07.110484 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:07 crc kubenswrapper[4861]: I0315 00:29:07.196851 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:09 crc kubenswrapper[4861]: I0315 00:29:09.181966 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "ba3891b9-a262-4c34-93bf-fd129e5e0b1d" (UID: "ba3891b9-a262-4c34-93bf-fd129e5e0b1d"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:09 crc kubenswrapper[4861]: I0315 00:29:09.234149 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/ba3891b9-a262-4c34-93bf-fd129e5e0b1d-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.780825 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:10 crc kubenswrapper[4861]: E0315 00:29:10.781703 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1ef5f4c-e426-43d2-90f8-c21be80ece91" containerName="oc" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781723 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1ef5f4c-e426-43d2-90f8-c21be80ece91" containerName="oc" Mar 15 00:29:10 crc kubenswrapper[4861]: E0315 00:29:10.781739 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="git-clone" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781748 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="git-clone" Mar 15 00:29:10 crc kubenswrapper[4861]: E0315 00:29:10.781768 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="manage-dockerfile" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781780 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="manage-dockerfile" Mar 15 00:29:10 crc kubenswrapper[4861]: E0315 00:29:10.781796 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="docker-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781806 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="docker-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781965 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1ef5f4c-e426-43d2-90f8-c21be80ece91" containerName="oc" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.781983 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba3891b9-a262-4c34-93bf-fd129e5e0b1d" containerName="docker-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.782896 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.786361 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.786737 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.787161 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.787210 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.811008 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.882954 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883046 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883201 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883264 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883307 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883331 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883362 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883437 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883516 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883591 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883630 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.883772 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgrcm\" (UniqueName: \"kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.985876 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgrcm\" (UniqueName: \"kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.985951 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.985980 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986034 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986064 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986103 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986134 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986297 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986845 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.986900 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.987017 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.987117 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.987253 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.987273 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.987396 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.988097 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.988138 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.988232 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.988338 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.988460 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.989134 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.996105 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:10 crc kubenswrapper[4861]: I0315 00:29:10.996315 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:11 crc kubenswrapper[4861]: I0315 00:29:11.019274 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgrcm\" (UniqueName: \"kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm\") pod \"sg-bridge-1-build\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:11 crc kubenswrapper[4861]: I0315 00:29:11.100398 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:11 crc kubenswrapper[4861]: I0315 00:29:11.580377 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:11 crc kubenswrapper[4861]: I0315 00:29:11.974339 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"b49f1a39-cd70-4c95-9fd2-5af26cba86c5","Type":"ContainerStarted","Data":"2aff0cf64d380cae16826edb0a51006f1b3c5cbd6ae0e6e19c988e624b958c17"} Mar 15 00:29:12 crc kubenswrapper[4861]: I0315 00:29:12.989074 4861 generic.go:334] "Generic (PLEG): container finished" podID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerID="e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b" exitCode=0 Mar 15 00:29:12 crc kubenswrapper[4861]: I0315 00:29:12.989149 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"b49f1a39-cd70-4c95-9fd2-5af26cba86c5","Type":"ContainerDied","Data":"e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b"} Mar 15 00:29:13 crc kubenswrapper[4861]: I0315 00:29:13.999281 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"b49f1a39-cd70-4c95-9fd2-5af26cba86c5","Type":"ContainerStarted","Data":"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b"} Mar 15 00:29:14 crc kubenswrapper[4861]: I0315 00:29:14.030633 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=4.030600819 podStartE2EDuration="4.030600819s" podCreationTimestamp="2026-03-15 00:29:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:29:14.027903435 +0000 UTC m=+1347.340101008" watchObservedRunningTime="2026-03-15 00:29:14.030600819 +0000 UTC m=+1347.342798392" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.257251 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.258645 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="docker-build" containerID="cri-o://197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b" gracePeriod=30 Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.665063 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_b49f1a39-cd70-4c95-9fd2-5af26cba86c5/docker-build/0.log" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.665719 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766211 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766667 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766695 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766755 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgrcm\" (UniqueName: \"kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766791 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766800 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766871 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766911 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766937 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.766973 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.767009 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.767123 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.767213 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.767330 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.767317 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push\") pod \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\" (UID: \"b49f1a39-cd70-4c95-9fd2-5af26cba86c5\") " Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768283 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768322 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768347 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768490 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768521 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768586 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.768790 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.773592 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm" (OuterVolumeSpecName: "kube-api-access-fgrcm") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "kube-api-access-fgrcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.774762 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.774788 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.836239 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870030 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870067 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870079 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870090 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870100 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870110 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870120 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:21 crc kubenswrapper[4861]: I0315 00:29:21.870130 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgrcm\" (UniqueName: \"kubernetes.io/projected/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-kube-api-access-fgrcm\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066230 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_b49f1a39-cd70-4c95-9fd2-5af26cba86c5/docker-build/0.log" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066761 4861 generic.go:334] "Generic (PLEG): container finished" podID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerID="197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b" exitCode=1 Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066804 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"b49f1a39-cd70-4c95-9fd2-5af26cba86c5","Type":"ContainerDied","Data":"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b"} Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066832 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"b49f1a39-cd70-4c95-9fd2-5af26cba86c5","Type":"ContainerDied","Data":"2aff0cf64d380cae16826edb0a51006f1b3c5cbd6ae0e6e19c988e624b958c17"} Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066851 4861 scope.go:117] "RemoveContainer" containerID="197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.066856 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.090109 4861 scope.go:117] "RemoveContainer" containerID="e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.117526 4861 scope.go:117] "RemoveContainer" containerID="197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b" Mar 15 00:29:22 crc kubenswrapper[4861]: E0315 00:29:22.118042 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b\": container with ID starting with 197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b not found: ID does not exist" containerID="197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.118106 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b"} err="failed to get container status \"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b\": rpc error: code = NotFound desc = could not find container \"197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b\": container with ID starting with 197efe5cef0c2b91d7794db765deae053dca4b47e154e8f42cfc891452445d2b not found: ID does not exist" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.118147 4861 scope.go:117] "RemoveContainer" containerID="e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b" Mar 15 00:29:22 crc kubenswrapper[4861]: E0315 00:29:22.118598 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b\": container with ID starting with e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b not found: ID does not exist" containerID="e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.118641 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b"} err="failed to get container status \"e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b\": rpc error: code = NotFound desc = could not find container \"e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b\": container with ID starting with e70dabd4f6e3176add5ce15e43bbb2290217ccd50cc10c3945ee6f665943ca3b not found: ID does not exist" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.217826 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "b49f1a39-cd70-4c95-9fd2-5af26cba86c5" (UID: "b49f1a39-cd70-4c95-9fd2-5af26cba86c5"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.276817 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b49f1a39-cd70-4c95-9fd2-5af26cba86c5-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.443250 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.463218 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.849947 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 15 00:29:22 crc kubenswrapper[4861]: E0315 00:29:22.850342 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="docker-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.850377 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="docker-build" Mar 15 00:29:22 crc kubenswrapper[4861]: E0315 00:29:22.850411 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="manage-dockerfile" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.850480 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="manage-dockerfile" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.850739 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" containerName="docker-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.852610 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.854779 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.855520 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.857029 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.859647 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.871989 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888426 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888515 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888593 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888715 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888848 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.888894 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889148 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889309 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889447 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmh5r\" (UniqueName: \"kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889539 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889673 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.889759 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991552 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991648 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991669 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991694 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991720 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991740 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991777 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991812 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991849 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmh5r\" (UniqueName: \"kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991875 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991898 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.991922 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.992528 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.992837 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993102 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993393 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993420 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993530 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993547 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.993661 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.994439 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.996902 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:22 crc kubenswrapper[4861]: I0315 00:29:22.998948 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:23 crc kubenswrapper[4861]: I0315 00:29:23.022892 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmh5r\" (UniqueName: \"kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r\") pod \"sg-bridge-2-build\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:23 crc kubenswrapper[4861]: I0315 00:29:23.175214 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 15 00:29:23 crc kubenswrapper[4861]: I0315 00:29:23.419816 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b49f1a39-cd70-4c95-9fd2-5af26cba86c5" path="/var/lib/kubelet/pods/b49f1a39-cd70-4c95-9fd2-5af26cba86c5/volumes" Mar 15 00:29:23 crc kubenswrapper[4861]: I0315 00:29:23.465304 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Mar 15 00:29:24 crc kubenswrapper[4861]: I0315 00:29:24.083961 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerStarted","Data":"6101d3eadc48dadc309734032009c5ed04f425082498b2b7fe30423ceba9c3b8"} Mar 15 00:29:24 crc kubenswrapper[4861]: I0315 00:29:24.084388 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerStarted","Data":"69b9dee48058b6c2463a160246a91820b4b0ae1ea37da9775448ffca50e84b30"} Mar 15 00:29:24 crc kubenswrapper[4861]: E0315 00:29:24.218003 4861 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.174:58614->38.102.83.174:45953: read tcp 38.102.83.174:58614->38.102.83.174:45953: read: connection reset by peer Mar 15 00:29:24 crc kubenswrapper[4861]: E0315 00:29:24.218131 4861 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:58614->38.102.83.174:45953: write tcp 38.102.83.174:58614->38.102.83.174:45953: write: broken pipe Mar 15 00:29:25 crc kubenswrapper[4861]: I0315 00:29:25.097484 4861 generic.go:334] "Generic (PLEG): container finished" podID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerID="6101d3eadc48dadc309734032009c5ed04f425082498b2b7fe30423ceba9c3b8" exitCode=0 Mar 15 00:29:25 crc kubenswrapper[4861]: I0315 00:29:25.097744 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerDied","Data":"6101d3eadc48dadc309734032009c5ed04f425082498b2b7fe30423ceba9c3b8"} Mar 15 00:29:26 crc kubenswrapper[4861]: I0315 00:29:26.110082 4861 generic.go:334] "Generic (PLEG): container finished" podID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerID="e4c4ed87fbc2a125d1fecf747b8fcf9e82aefd85c3d6a74a651e280025c11ce3" exitCode=0 Mar 15 00:29:26 crc kubenswrapper[4861]: I0315 00:29:26.110148 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerDied","Data":"e4c4ed87fbc2a125d1fecf747b8fcf9e82aefd85c3d6a74a651e280025c11ce3"} Mar 15 00:29:26 crc kubenswrapper[4861]: I0315 00:29:26.175268 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_00e87192-226f-4d84-a424-d4adf2ca07e1/manage-dockerfile/0.log" Mar 15 00:29:27 crc kubenswrapper[4861]: I0315 00:29:27.121240 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerStarted","Data":"8d33c03fdacf9dff5bf6b41713c47bd42526f7945b40f8c2394261ebf239e4bd"} Mar 15 00:29:27 crc kubenswrapper[4861]: I0315 00:29:27.212315 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.212293834 podStartE2EDuration="5.212293834s" podCreationTimestamp="2026-03-15 00:29:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:29:27.205890078 +0000 UTC m=+1360.518087621" watchObservedRunningTime="2026-03-15 00:29:27.212293834 +0000 UTC m=+1360.524491387" Mar 15 00:29:32 crc kubenswrapper[4861]: I0315 00:29:32.282855 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:29:32 crc kubenswrapper[4861]: I0315 00:29:32.284042 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.163200 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv"] Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.165250 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.167516 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.168478 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.171290 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558910-gjt28"] Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.173090 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.177067 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558910-gjt28"] Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.179839 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.180264 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.182689 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv"] Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.184041 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.297583 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.297674 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.297698 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mgzw\" (UniqueName: \"kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.297763 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj9w5\" (UniqueName: \"kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5\") pod \"auto-csr-approver-29558910-gjt28\" (UID: \"89f5a73f-6841-4b40-be0c-31c0e4760f86\") " pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.399406 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj9w5\" (UniqueName: \"kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5\") pod \"auto-csr-approver-29558910-gjt28\" (UID: \"89f5a73f-6841-4b40-be0c-31c0e4760f86\") " pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.399519 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.399626 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.399646 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mgzw\" (UniqueName: \"kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.401122 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.408269 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.421793 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mgzw\" (UniqueName: \"kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw\") pod \"collect-profiles-29558910-jhtzv\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.422523 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj9w5\" (UniqueName: \"kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5\") pod \"auto-csr-approver-29558910-gjt28\" (UID: \"89f5a73f-6841-4b40-be0c-31c0e4760f86\") " pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.501609 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.514221 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.759232 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv"] Mar 15 00:30:00 crc kubenswrapper[4861]: W0315 00:30:00.812045 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89f5a73f_6841_4b40_be0c_31c0e4760f86.slice/crio-6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e WatchSource:0}: Error finding container 6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e: Status 404 returned error can't find the container with id 6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e Mar 15 00:30:00 crc kubenswrapper[4861]: I0315 00:30:00.812602 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558910-gjt28"] Mar 15 00:30:01 crc kubenswrapper[4861]: I0315 00:30:01.423682 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558910-gjt28" event={"ID":"89f5a73f-6841-4b40-be0c-31c0e4760f86","Type":"ContainerStarted","Data":"6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e"} Mar 15 00:30:01 crc kubenswrapper[4861]: I0315 00:30:01.426559 4861 generic.go:334] "Generic (PLEG): container finished" podID="57eaf526-5a8c-4b81-9109-64a115929c17" containerID="fba44dc6ffa1f7692405296ccbdcc7503622298da11949a3c174ec812ae53673" exitCode=0 Mar 15 00:30:01 crc kubenswrapper[4861]: I0315 00:30:01.426644 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" event={"ID":"57eaf526-5a8c-4b81-9109-64a115929c17","Type":"ContainerDied","Data":"fba44dc6ffa1f7692405296ccbdcc7503622298da11949a3c174ec812ae53673"} Mar 15 00:30:01 crc kubenswrapper[4861]: I0315 00:30:01.426667 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" event={"ID":"57eaf526-5a8c-4b81-9109-64a115929c17","Type":"ContainerStarted","Data":"41b357e8cb11d568d2b291176aa8571fd5e853e6aa7ad7476b5ac6bbd8259fd6"} Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.282300 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.282829 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.438014 4861 generic.go:334] "Generic (PLEG): container finished" podID="89f5a73f-6841-4b40-be0c-31c0e4760f86" containerID="c80cf9f3d8c68c3730133799c00a540ce5a01f1d6d6e4a710cde1ba836f2e2ce" exitCode=0 Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.438090 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558910-gjt28" event={"ID":"89f5a73f-6841-4b40-be0c-31c0e4760f86","Type":"ContainerDied","Data":"c80cf9f3d8c68c3730133799c00a540ce5a01f1d6d6e4a710cde1ba836f2e2ce"} Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.688974 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.736169 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume\") pod \"57eaf526-5a8c-4b81-9109-64a115929c17\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.736237 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mgzw\" (UniqueName: \"kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw\") pod \"57eaf526-5a8c-4b81-9109-64a115929c17\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.736397 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume\") pod \"57eaf526-5a8c-4b81-9109-64a115929c17\" (UID: \"57eaf526-5a8c-4b81-9109-64a115929c17\") " Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.737417 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume" (OuterVolumeSpecName: "config-volume") pod "57eaf526-5a8c-4b81-9109-64a115929c17" (UID: "57eaf526-5a8c-4b81-9109-64a115929c17"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.743252 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "57eaf526-5a8c-4b81-9109-64a115929c17" (UID: "57eaf526-5a8c-4b81-9109-64a115929c17"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.746688 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw" (OuterVolumeSpecName: "kube-api-access-6mgzw") pod "57eaf526-5a8c-4b81-9109-64a115929c17" (UID: "57eaf526-5a8c-4b81-9109-64a115929c17"). InnerVolumeSpecName "kube-api-access-6mgzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.838383 4861 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/57eaf526-5a8c-4b81-9109-64a115929c17-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.838434 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mgzw\" (UniqueName: \"kubernetes.io/projected/57eaf526-5a8c-4b81-9109-64a115929c17-kube-api-access-6mgzw\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:02 crc kubenswrapper[4861]: I0315 00:30:02.838611 4861 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/57eaf526-5a8c-4b81-9109-64a115929c17-config-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.453361 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.453640 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29558910-jhtzv" event={"ID":"57eaf526-5a8c-4b81-9109-64a115929c17","Type":"ContainerDied","Data":"41b357e8cb11d568d2b291176aa8571fd5e853e6aa7ad7476b5ac6bbd8259fd6"} Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.453739 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41b357e8cb11d568d2b291176aa8571fd5e853e6aa7ad7476b5ac6bbd8259fd6" Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.741283 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.855006 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj9w5\" (UniqueName: \"kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5\") pod \"89f5a73f-6841-4b40-be0c-31c0e4760f86\" (UID: \"89f5a73f-6841-4b40-be0c-31c0e4760f86\") " Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.866636 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5" (OuterVolumeSpecName: "kube-api-access-vj9w5") pod "89f5a73f-6841-4b40-be0c-31c0e4760f86" (UID: "89f5a73f-6841-4b40-be0c-31c0e4760f86"). InnerVolumeSpecName "kube-api-access-vj9w5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:30:03 crc kubenswrapper[4861]: I0315 00:30:03.956883 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj9w5\" (UniqueName: \"kubernetes.io/projected/89f5a73f-6841-4b40-be0c-31c0e4760f86-kube-api-access-vj9w5\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:04 crc kubenswrapper[4861]: I0315 00:30:04.461375 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558910-gjt28" event={"ID":"89f5a73f-6841-4b40-be0c-31c0e4760f86","Type":"ContainerDied","Data":"6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e"} Mar 15 00:30:04 crc kubenswrapper[4861]: I0315 00:30:04.461434 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cefe851ae560eaab4077ad2139c503c22d86ec9568e4c19de0837ad3d650a9e" Mar 15 00:30:04 crc kubenswrapper[4861]: I0315 00:30:04.461523 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558910-gjt28" Mar 15 00:30:04 crc kubenswrapper[4861]: I0315 00:30:04.826083 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558904-b8spq"] Mar 15 00:30:04 crc kubenswrapper[4861]: I0315 00:30:04.831981 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558904-b8spq"] Mar 15 00:30:05 crc kubenswrapper[4861]: I0315 00:30:05.421842 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6e6694-d808-442b-b2af-e3d059b3a779" path="/var/lib/kubelet/pods/8d6e6694-d808-442b-b2af-e3d059b3a779/volumes" Mar 15 00:30:16 crc kubenswrapper[4861]: I0315 00:30:16.570444 4861 generic.go:334] "Generic (PLEG): container finished" podID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerID="8d33c03fdacf9dff5bf6b41713c47bd42526f7945b40f8c2394261ebf239e4bd" exitCode=0 Mar 15 00:30:16 crc kubenswrapper[4861]: I0315 00:30:16.570714 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerDied","Data":"8d33c03fdacf9dff5bf6b41713c47bd42526f7945b40f8c2394261ebf239e4bd"} Mar 15 00:30:17 crc kubenswrapper[4861]: I0315 00:30:17.896310 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.072867 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmh5r\" (UniqueName: \"kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073060 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073213 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073272 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073380 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073443 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073490 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073708 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073794 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073854 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.073942 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.074020 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push\") pod \"00e87192-226f-4d84-a424-d4adf2ca07e1\" (UID: \"00e87192-226f-4d84-a424-d4adf2ca07e1\") " Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.074153 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.074698 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.074768 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.075049 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.075110 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.077274 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.077492 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.081625 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r" (OuterVolumeSpecName: "kube-api-access-dmh5r") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "kube-api-access-dmh5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.083290 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.083462 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.083543 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175172 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175277 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175290 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175302 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmh5r\" (UniqueName: \"kubernetes.io/projected/00e87192-226f-4d84-a424-d4adf2ca07e1-kube-api-access-dmh5r\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175312 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175322 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175330 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/00e87192-226f-4d84-a424-d4adf2ca07e1-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175339 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/00e87192-226f-4d84-a424-d4adf2ca07e1-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.175348 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/00e87192-226f-4d84-a424-d4adf2ca07e1-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.217529 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.278527 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.595994 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"00e87192-226f-4d84-a424-d4adf2ca07e1","Type":"ContainerDied","Data":"69b9dee48058b6c2463a160246a91820b4b0ae1ea37da9775448ffca50e84b30"} Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.596050 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69b9dee48058b6c2463a160246a91820b4b0ae1ea37da9775448ffca50e84b30" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.596151 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.949859 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "00e87192-226f-4d84-a424-d4adf2ca07e1" (UID: "00e87192-226f-4d84-a424-d4adf2ca07e1"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:18 crc kubenswrapper[4861]: I0315 00:30:18.991073 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/00e87192-226f-4d84-a424-d4adf2ca07e1-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.969036 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:21 crc kubenswrapper[4861]: E0315 00:30:21.971447 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="manage-dockerfile" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.971677 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="manage-dockerfile" Mar 15 00:30:21 crc kubenswrapper[4861]: E0315 00:30:21.971815 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89f5a73f-6841-4b40-be0c-31c0e4760f86" containerName="oc" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.971960 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="89f5a73f-6841-4b40-be0c-31c0e4760f86" containerName="oc" Mar 15 00:30:21 crc kubenswrapper[4861]: E0315 00:30:21.972119 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="docker-build" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.972257 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="docker-build" Mar 15 00:30:21 crc kubenswrapper[4861]: E0315 00:30:21.972389 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="git-clone" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.972506 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="git-clone" Mar 15 00:30:21 crc kubenswrapper[4861]: E0315 00:30:21.972698 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57eaf526-5a8c-4b81-9109-64a115929c17" containerName="collect-profiles" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.972824 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="57eaf526-5a8c-4b81-9109-64a115929c17" containerName="collect-profiles" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.973164 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="89f5a73f-6841-4b40-be0c-31c0e4760f86" containerName="oc" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.973329 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e87192-226f-4d84-a424-d4adf2ca07e1" containerName="docker-build" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.973519 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="57eaf526-5a8c-4b81-9109-64a115929c17" containerName="collect-profiles" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.976838 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.981695 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.982043 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.982211 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.982374 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Mar 15 00:30:21 crc kubenswrapper[4861]: I0315 00:30:21.993442 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045206 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045257 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045284 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045317 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045350 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045377 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045433 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045469 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045498 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbg8x\" (UniqueName: \"kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045516 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045536 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.045622 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146234 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146303 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146325 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146348 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146383 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146411 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146435 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146471 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146494 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146519 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbg8x\" (UniqueName: \"kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146545 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146585 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146807 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.146891 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.147096 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.147148 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.147768 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.148348 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.148389 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.148710 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.149249 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.155590 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.155595 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.166162 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbg8x\" (UniqueName: \"kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.297045 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:22 crc kubenswrapper[4861]: I0315 00:30:22.818153 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:23 crc kubenswrapper[4861]: I0315 00:30:23.648645 4861 generic.go:334] "Generic (PLEG): container finished" podID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerID="360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5" exitCode=0 Mar 15 00:30:23 crc kubenswrapper[4861]: I0315 00:30:23.648896 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"112fac0b-a393-43eb-9931-2be8bfa8560c","Type":"ContainerDied","Data":"360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5"} Mar 15 00:30:23 crc kubenswrapper[4861]: I0315 00:30:23.649215 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"112fac0b-a393-43eb-9931-2be8bfa8560c","Type":"ContainerStarted","Data":"1713a2cf181cf16c4e5159d9e2c154b3be686004c2adab7b20937c2e085fe858"} Mar 15 00:30:24 crc kubenswrapper[4861]: I0315 00:30:24.661018 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"112fac0b-a393-43eb-9931-2be8bfa8560c","Type":"ContainerStarted","Data":"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3"} Mar 15 00:30:24 crc kubenswrapper[4861]: I0315 00:30:24.706773 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=3.706743223 podStartE2EDuration="3.706743223s" podCreationTimestamp="2026-03-15 00:30:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:30:24.69814141 +0000 UTC m=+1418.010338983" watchObservedRunningTime="2026-03-15 00:30:24.706743223 +0000 UTC m=+1418.018940796" Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.282677 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.283274 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.283360 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.284628 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.284735 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19" gracePeriod=600 Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.711640 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.712350 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="docker-build" containerID="cri-o://963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3" gracePeriod=30 Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.746147 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19" exitCode=0 Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.746231 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19"} Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.746297 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187"} Mar 15 00:30:32 crc kubenswrapper[4861]: I0315 00:30:32.746334 4861 scope.go:117] "RemoveContainer" containerID="d87a74b3db9446d2fd4b465725e02497b490db70eabc3874bf6146dd5354c267" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.134985 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_112fac0b-a393-43eb-9931-2be8bfa8560c/docker-build/0.log" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.135980 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.331684 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332097 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332141 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbg8x\" (UniqueName: \"kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332181 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332219 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332239 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332265 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332337 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332360 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332402 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332447 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332463 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets\") pod \"112fac0b-a393-43eb-9931-2be8bfa8560c\" (UID: \"112fac0b-a393-43eb-9931-2be8bfa8560c\") " Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332742 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.332938 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.333601 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.333661 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.333863 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.333961 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.334224 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.339468 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.339501 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.339561 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x" (OuterVolumeSpecName: "kube-api-access-rbg8x") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "kube-api-access-rbg8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.427831 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444707 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444753 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/112fac0b-a393-43eb-9931-2be8bfa8560c-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444771 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444786 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444800 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbg8x\" (UniqueName: \"kubernetes.io/projected/112fac0b-a393-43eb-9931-2be8bfa8560c-kube-api-access-rbg8x\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444819 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444833 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/112fac0b-a393-43eb-9931-2be8bfa8560c-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444846 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444861 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444876 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.444888 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/112fac0b-a393-43eb-9931-2be8bfa8560c-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.758162 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_112fac0b-a393-43eb-9931-2be8bfa8560c/docker-build/0.log" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.758893 4861 generic.go:334] "Generic (PLEG): container finished" podID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerID="963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3" exitCode=1 Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.758993 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.759028 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"112fac0b-a393-43eb-9931-2be8bfa8560c","Type":"ContainerDied","Data":"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3"} Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.759122 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"112fac0b-a393-43eb-9931-2be8bfa8560c","Type":"ContainerDied","Data":"1713a2cf181cf16c4e5159d9e2c154b3be686004c2adab7b20937c2e085fe858"} Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.759163 4861 scope.go:117] "RemoveContainer" containerID="963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.779748 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "112fac0b-a393-43eb-9931-2be8bfa8560c" (UID: "112fac0b-a393-43eb-9931-2be8bfa8560c"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.791980 4861 scope.go:117] "RemoveContainer" containerID="360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.827657 4861 scope.go:117] "RemoveContainer" containerID="963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3" Mar 15 00:30:33 crc kubenswrapper[4861]: E0315 00:30:33.828431 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3\": container with ID starting with 963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3 not found: ID does not exist" containerID="963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.828753 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3"} err="failed to get container status \"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3\": rpc error: code = NotFound desc = could not find container \"963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3\": container with ID starting with 963b6473dbdbd3d4c75492dabb10246ad6a7c1dc5f6d7588060e6a0deff283e3 not found: ID does not exist" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.828945 4861 scope.go:117] "RemoveContainer" containerID="360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5" Mar 15 00:30:33 crc kubenswrapper[4861]: E0315 00:30:33.829828 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5\": container with ID starting with 360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5 not found: ID does not exist" containerID="360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.829917 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5"} err="failed to get container status \"360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5\": rpc error: code = NotFound desc = could not find container \"360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5\": container with ID starting with 360643d94b1ab3628da89d7612142569a2ed2ccca2169006baa6416cc1bf22c5 not found: ID does not exist" Mar 15 00:30:33 crc kubenswrapper[4861]: I0315 00:30:33.850915 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/112fac0b-a393-43eb-9931-2be8bfa8560c-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.118174 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.125272 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.296991 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 15 00:30:34 crc kubenswrapper[4861]: E0315 00:30:34.297604 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="docker-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.297621 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="docker-build" Mar 15 00:30:34 crc kubenswrapper[4861]: E0315 00:30:34.297643 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="manage-dockerfile" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.297652 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="manage-dockerfile" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.297809 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" containerName="docker-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.298954 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.301345 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.301953 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.301958 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.303814 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.329383 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.359874 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360070 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360232 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxp85\" (UniqueName: \"kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360299 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360386 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360424 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360444 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360508 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360604 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360629 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360655 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.360744 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.462979 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463053 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463081 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463128 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463162 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463181 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463204 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463235 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.463510 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.465335 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.465432 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.465641 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxp85\" (UniqueName: \"kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.465710 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.466163 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.467881 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.467991 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.468018 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.468266 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.468273 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.472895 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.474054 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.475291 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.480305 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.485114 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxp85\" (UniqueName: \"kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.647535 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:30:34 crc kubenswrapper[4861]: I0315 00:30:34.957675 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Mar 15 00:30:35 crc kubenswrapper[4861]: I0315 00:30:35.417245 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="112fac0b-a393-43eb-9931-2be8bfa8560c" path="/var/lib/kubelet/pods/112fac0b-a393-43eb-9931-2be8bfa8560c/volumes" Mar 15 00:30:35 crc kubenswrapper[4861]: I0315 00:30:35.798606 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerStarted","Data":"cf00e795b6adab279612ccaaea688f5b352e5593e89585203e5ce6f7fc212c13"} Mar 15 00:30:35 crc kubenswrapper[4861]: I0315 00:30:35.798944 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerStarted","Data":"cbdce4b2a3400aadd3d26e6370a511a0b9283634c5c39f0fc4fde0d513606232"} Mar 15 00:30:36 crc kubenswrapper[4861]: I0315 00:30:36.809606 4861 generic.go:334] "Generic (PLEG): container finished" podID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerID="cf00e795b6adab279612ccaaea688f5b352e5593e89585203e5ce6f7fc212c13" exitCode=0 Mar 15 00:30:36 crc kubenswrapper[4861]: I0315 00:30:36.809695 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerDied","Data":"cf00e795b6adab279612ccaaea688f5b352e5593e89585203e5ce6f7fc212c13"} Mar 15 00:30:37 crc kubenswrapper[4861]: I0315 00:30:37.823679 4861 generic.go:334] "Generic (PLEG): container finished" podID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerID="15f85dfb20ef8b44ec629286568c92d3f04eb56cdeb3cbf16e7128c9727f2b83" exitCode=0 Mar 15 00:30:37 crc kubenswrapper[4861]: I0315 00:30:37.823810 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerDied","Data":"15f85dfb20ef8b44ec629286568c92d3f04eb56cdeb3cbf16e7128c9727f2b83"} Mar 15 00:30:37 crc kubenswrapper[4861]: I0315 00:30:37.893985 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_64eaebbd-31b2-4ea3-9192-3d807d8edbf0/manage-dockerfile/0.log" Mar 15 00:30:38 crc kubenswrapper[4861]: I0315 00:30:38.835488 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerStarted","Data":"df78f3acfbc1409e074e00b3c47238ac2b54da3cd9db84e105f1b991dc933c02"} Mar 15 00:30:38 crc kubenswrapper[4861]: I0315 00:30:38.874160 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=4.874128442 podStartE2EDuration="4.874128442s" podCreationTimestamp="2026-03-15 00:30:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:30:38.870951616 +0000 UTC m=+1432.183149179" watchObservedRunningTime="2026-03-15 00:30:38.874128442 +0000 UTC m=+1432.186326015" Mar 15 00:30:48 crc kubenswrapper[4861]: I0315 00:30:48.564723 4861 scope.go:117] "RemoveContainer" containerID="4e65d81d583359399aac371fdf4d6abe208f4e396a277b6b3e769b9dabdf7874" Mar 15 00:31:32 crc kubenswrapper[4861]: I0315 00:31:32.271737 4861 generic.go:334] "Generic (PLEG): container finished" podID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerID="df78f3acfbc1409e074e00b3c47238ac2b54da3cd9db84e105f1b991dc933c02" exitCode=0 Mar 15 00:31:32 crc kubenswrapper[4861]: I0315 00:31:32.271941 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerDied","Data":"df78f3acfbc1409e074e00b3c47238ac2b54da3cd9db84e105f1b991dc933c02"} Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.560709 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712294 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712393 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712422 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712456 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxp85\" (UniqueName: \"kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712498 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712580 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712616 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712652 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712696 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712738 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712772 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.712798 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run\") pod \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\" (UID: \"64eaebbd-31b2-4ea3-9192-3d807d8edbf0\") " Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.713404 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.713690 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.714372 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.714861 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.714923 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.717013 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.717384 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.720919 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.721754 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.722865 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85" (OuterVolumeSpecName: "kube-api-access-vxp85") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "kube-api-access-vxp85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815089 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815135 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815144 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815155 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815165 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815174 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815182 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxp85\" (UniqueName: \"kubernetes.io/projected/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-kube-api-access-vxp85\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815192 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815200 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.815211 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.855206 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:33 crc kubenswrapper[4861]: I0315 00:31:33.916366 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:34 crc kubenswrapper[4861]: I0315 00:31:34.289578 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"64eaebbd-31b2-4ea3-9192-3d807d8edbf0","Type":"ContainerDied","Data":"cbdce4b2a3400aadd3d26e6370a511a0b9283634c5c39f0fc4fde0d513606232"} Mar 15 00:31:34 crc kubenswrapper[4861]: I0315 00:31:34.289907 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbdce4b2a3400aadd3d26e6370a511a0b9283634c5c39f0fc4fde0d513606232" Mar 15 00:31:34 crc kubenswrapper[4861]: I0315 00:31:34.289687 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Mar 15 00:31:34 crc kubenswrapper[4861]: I0315 00:31:34.674839 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "64eaebbd-31b2-4ea3-9192-3d807d8edbf0" (UID: "64eaebbd-31b2-4ea3-9192-3d807d8edbf0"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:34 crc kubenswrapper[4861]: I0315 00:31:34.729920 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/64eaebbd-31b2-4ea3-9192-3d807d8edbf0-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.605561 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Mar 15 00:31:43 crc kubenswrapper[4861]: E0315 00:31:43.607053 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="docker-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.607084 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="docker-build" Mar 15 00:31:43 crc kubenswrapper[4861]: E0315 00:31:43.607128 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="manage-dockerfile" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.607146 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="manage-dockerfile" Mar 15 00:31:43 crc kubenswrapper[4861]: E0315 00:31:43.607182 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="git-clone" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.607198 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="git-clone" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.607529 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="64eaebbd-31b2-4ea3-9192-3d807d8edbf0" containerName="docker-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.609266 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.612283 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-global-ca" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.615183 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-sys-config" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.616340 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-1-ca" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.616356 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.641127 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677722 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677783 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677852 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677893 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677926 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.677960 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678015 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678052 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678080 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678109 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4xhw\" (UniqueName: \"kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678132 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.678168 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780262 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780368 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780411 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780461 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780506 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780550 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780641 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780675 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780786 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780867 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780915 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.780980 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781014 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4xhw\" (UniqueName: \"kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781088 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781407 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781482 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781602 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781767 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.781860 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.782862 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.783452 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.791634 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.791642 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.812802 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4xhw\" (UniqueName: \"kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw\") pod \"service-telemetry-operator-bundle-1-build\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:43 crc kubenswrapper[4861]: I0315 00:31:43.944218 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:44 crc kubenswrapper[4861]: I0315 00:31:44.270800 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Mar 15 00:31:44 crc kubenswrapper[4861]: I0315 00:31:44.380079 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"6d0c36e0-9af2-4b74-b563-1dbdc911f854","Type":"ContainerStarted","Data":"ce0827d3a2eef8c899c12ddd5aef362b75d40505711d1fce9e8413e20b99836d"} Mar 15 00:31:45 crc kubenswrapper[4861]: I0315 00:31:45.391578 4861 generic.go:334] "Generic (PLEG): container finished" podID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerID="45b662408f36b7f9e4a466d8b7d1ac0a72d03af330284ea8b47ad01cbad69089" exitCode=0 Mar 15 00:31:45 crc kubenswrapper[4861]: I0315 00:31:45.391669 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"6d0c36e0-9af2-4b74-b563-1dbdc911f854","Type":"ContainerDied","Data":"45b662408f36b7f9e4a466d8b7d1ac0a72d03af330284ea8b47ad01cbad69089"} Mar 15 00:31:46 crc kubenswrapper[4861]: I0315 00:31:46.405605 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_6d0c36e0-9af2-4b74-b563-1dbdc911f854/docker-build/0.log" Mar 15 00:31:46 crc kubenswrapper[4861]: I0315 00:31:46.407346 4861 generic.go:334] "Generic (PLEG): container finished" podID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerID="1bf4e329bde3aef2737a7208e9fa79e473602819bd693a8fca561f207be6d179" exitCode=1 Mar 15 00:31:46 crc kubenswrapper[4861]: I0315 00:31:46.407396 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"6d0c36e0-9af2-4b74-b563-1dbdc911f854","Type":"ContainerDied","Data":"1bf4e329bde3aef2737a7208e9fa79e473602819bd693a8fca561f207be6d179"} Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.727264 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_6d0c36e0-9af2-4b74-b563-1dbdc911f854/docker-build/0.log" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.728452 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.748793 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.748887 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.748926 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749055 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749090 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749005 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749114 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749108 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749502 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749588 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749659 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749702 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749782 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4xhw\" (UniqueName: \"kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749810 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root\") pod \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\" (UID: \"6d0c36e0-9af2-4b74-b563-1dbdc911f854\") " Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.749888 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.751334 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.751361 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.752428 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.753170 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.753546 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.754038 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.755040 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.756017 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.758850 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.758859 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw" (OuterVolumeSpecName: "kube-api-access-h4xhw") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "kube-api-access-h4xhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.759789 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "6d0c36e0-9af2-4b74-b563-1dbdc911f854" (UID: "6d0c36e0-9af2-4b74-b563-1dbdc911f854"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.852400 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4xhw\" (UniqueName: \"kubernetes.io/projected/6d0c36e0-9af2-4b74-b563-1dbdc911f854-kube-api-access-h4xhw\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.852834 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.852976 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853109 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853275 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853424 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853552 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/6d0c36e0-9af2-4b74-b563-1dbdc911f854-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853746 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.853875 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d0c36e0-9af2-4b74-b563-1dbdc911f854-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:47 crc kubenswrapper[4861]: I0315 00:31:47.854020 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d0c36e0-9af2-4b74-b563-1dbdc911f854-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:31:48 crc kubenswrapper[4861]: I0315 00:31:48.433129 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-1-build_6d0c36e0-9af2-4b74-b563-1dbdc911f854/docker-build/0.log" Mar 15 00:31:48 crc kubenswrapper[4861]: I0315 00:31:48.433699 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-1-build" event={"ID":"6d0c36e0-9af2-4b74-b563-1dbdc911f854","Type":"ContainerDied","Data":"ce0827d3a2eef8c899c12ddd5aef362b75d40505711d1fce9e8413e20b99836d"} Mar 15 00:31:48 crc kubenswrapper[4861]: I0315 00:31:48.433759 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce0827d3a2eef8c899c12ddd5aef362b75d40505711d1fce9e8413e20b99836d" Mar 15 00:31:48 crc kubenswrapper[4861]: I0315 00:31:48.433789 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-1-build" Mar 15 00:31:54 crc kubenswrapper[4861]: I0315 00:31:54.026917 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Mar 15 00:31:54 crc kubenswrapper[4861]: I0315 00:31:54.037289 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-1-build"] Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.420423 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" path="/var/lib/kubelet/pods/6d0c36e0-9af2-4b74-b563-1dbdc911f854/volumes" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.778777 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Mar 15 00:31:55 crc kubenswrapper[4861]: E0315 00:31:55.783006 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerName="manage-dockerfile" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.783130 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerName="manage-dockerfile" Mar 15 00:31:55 crc kubenswrapper[4861]: E0315 00:31:55.783192 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerName="docker-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.783254 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerName="docker-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.783605 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d0c36e0-9af2-4b74-b563-1dbdc911f854" containerName="docker-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.789142 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.792495 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.794435 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-ca" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.794844 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-global-ca" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.795155 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-bundle-2-sys-config" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.887985 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.954170 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.954413 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.954496 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.954835 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.954902 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955016 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955043 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955151 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955202 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955252 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955272 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:55 crc kubenswrapper[4861]: I0315 00:31:55.955291 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6tsr\" (UniqueName: \"kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056728 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056813 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056844 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056869 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056892 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6tsr\" (UniqueName: \"kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056922 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056959 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.056989 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057048 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057071 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057100 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057126 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057336 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057440 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057545 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.057711 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.058100 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.058215 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.058253 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.058593 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.058883 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.065540 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.065988 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.086144 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6tsr\" (UniqueName: \"kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr\") pod \"service-telemetry-operator-bundle-2-build\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.191891 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.457249 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-bundle-2-build"] Mar 15 00:31:56 crc kubenswrapper[4861]: I0315 00:31:56.521679 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerStarted","Data":"36ddba62092a70f310bf7fe3397b72578834334d9af2792a2e445ddeaa5050b9"} Mar 15 00:31:57 crc kubenswrapper[4861]: I0315 00:31:57.532473 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerStarted","Data":"7c5364a1d5b72cc6ef8f58ef42a119f2a7f7863048f148fca0735fde1632e627"} Mar 15 00:31:58 crc kubenswrapper[4861]: I0315 00:31:58.543329 4861 generic.go:334] "Generic (PLEG): container finished" podID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerID="7c5364a1d5b72cc6ef8f58ef42a119f2a7f7863048f148fca0735fde1632e627" exitCode=0 Mar 15 00:31:58 crc kubenswrapper[4861]: I0315 00:31:58.543416 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerDied","Data":"7c5364a1d5b72cc6ef8f58ef42a119f2a7f7863048f148fca0735fde1632e627"} Mar 15 00:31:59 crc kubenswrapper[4861]: I0315 00:31:59.556216 4861 generic.go:334] "Generic (PLEG): container finished" podID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerID="6075d384349a5592dca024f0b66c4c3b28c45820405176c496ac12512dd55913" exitCode=0 Mar 15 00:31:59 crc kubenswrapper[4861]: I0315 00:31:59.556348 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerDied","Data":"6075d384349a5592dca024f0b66c4c3b28c45820405176c496ac12512dd55913"} Mar 15 00:31:59 crc kubenswrapper[4861]: I0315 00:31:59.630062 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-bundle-2-build_da2c16cb-5691-4679-8591-fdfdcee443e7/manage-dockerfile/0.log" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.147886 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558912-qwsn4"] Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.148820 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.151190 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.152704 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.152727 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.159573 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558912-qwsn4"] Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.336332 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxg85\" (UniqueName: \"kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85\") pod \"auto-csr-approver-29558912-qwsn4\" (UID: \"67bf6356-73f3-4534-84a0-6f5ea23d7c77\") " pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.439981 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxg85\" (UniqueName: \"kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85\") pod \"auto-csr-approver-29558912-qwsn4\" (UID: \"67bf6356-73f3-4534-84a0-6f5ea23d7c77\") " pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.468185 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxg85\" (UniqueName: \"kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85\") pod \"auto-csr-approver-29558912-qwsn4\" (UID: \"67bf6356-73f3-4534-84a0-6f5ea23d7c77\") " pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.570717 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerStarted","Data":"1c660172deb7b1cd5168af47c71f64bfd86d1b5c349386273f0759512b995655"} Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.612550 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-bundle-2-build" podStartSLOduration=5.6125271770000005 podStartE2EDuration="5.612527177s" podCreationTimestamp="2026-03-15 00:31:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:32:00.605816686 +0000 UTC m=+1513.918014229" watchObservedRunningTime="2026-03-15 00:32:00.612527177 +0000 UTC m=+1513.924724720" Mar 15 00:32:00 crc kubenswrapper[4861]: I0315 00:32:00.764962 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:01 crc kubenswrapper[4861]: I0315 00:32:01.019664 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558912-qwsn4"] Mar 15 00:32:01 crc kubenswrapper[4861]: W0315 00:32:01.029289 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67bf6356_73f3_4534_84a0_6f5ea23d7c77.slice/crio-dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4 WatchSource:0}: Error finding container dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4: Status 404 returned error can't find the container with id dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4 Mar 15 00:32:01 crc kubenswrapper[4861]: I0315 00:32:01.581753 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" event={"ID":"67bf6356-73f3-4534-84a0-6f5ea23d7c77","Type":"ContainerStarted","Data":"dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4"} Mar 15 00:32:02 crc kubenswrapper[4861]: I0315 00:32:02.595237 4861 generic.go:334] "Generic (PLEG): container finished" podID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerID="1c660172deb7b1cd5168af47c71f64bfd86d1b5c349386273f0759512b995655" exitCode=0 Mar 15 00:32:02 crc kubenswrapper[4861]: I0315 00:32:02.595770 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerDied","Data":"1c660172deb7b1cd5168af47c71f64bfd86d1b5c349386273f0759512b995655"} Mar 15 00:32:02 crc kubenswrapper[4861]: I0315 00:32:02.598399 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" event={"ID":"67bf6356-73f3-4534-84a0-6f5ea23d7c77","Type":"ContainerStarted","Data":"3ebd11baefc4351d7d192681acc127e4ed22ca5a6378f5b998144cf8cc6dffb3"} Mar 15 00:32:02 crc kubenswrapper[4861]: I0315 00:32:02.677526 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" podStartSLOduration=1.6975649769999999 podStartE2EDuration="2.677493923s" podCreationTimestamp="2026-03-15 00:32:00 +0000 UTC" firstStartedPulling="2026-03-15 00:32:01.033081919 +0000 UTC m=+1514.345279452" lastFinishedPulling="2026-03-15 00:32:02.013010865 +0000 UTC m=+1515.325208398" observedRunningTime="2026-03-15 00:32:02.675032506 +0000 UTC m=+1515.987230079" watchObservedRunningTime="2026-03-15 00:32:02.677493923 +0000 UTC m=+1515.989691456" Mar 15 00:32:03 crc kubenswrapper[4861]: I0315 00:32:03.611799 4861 generic.go:334] "Generic (PLEG): container finished" podID="67bf6356-73f3-4534-84a0-6f5ea23d7c77" containerID="3ebd11baefc4351d7d192681acc127e4ed22ca5a6378f5b998144cf8cc6dffb3" exitCode=0 Mar 15 00:32:03 crc kubenswrapper[4861]: I0315 00:32:03.611940 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" event={"ID":"67bf6356-73f3-4534-84a0-6f5ea23d7c77","Type":"ContainerDied","Data":"3ebd11baefc4351d7d192681acc127e4ed22ca5a6378f5b998144cf8cc6dffb3"} Mar 15 00:32:03 crc kubenswrapper[4861]: I0315 00:32:03.950720 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101591 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101702 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101770 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101810 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101822 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6tsr\" (UniqueName: \"kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101849 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101871 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.101969 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.102009 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.102041 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.102088 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.102121 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103132 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103197 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103281 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103483 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103524 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run\") pod \"da2c16cb-5691-4679-8591-fdfdcee443e7\" (UID: \"da2c16cb-5691-4679-8591-fdfdcee443e7\") " Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103876 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103901 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103918 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/da2c16cb-5691-4679-8591-fdfdcee443e7-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103935 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.103952 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.104397 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.104963 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.108323 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.109046 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.109540 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr" (OuterVolumeSpecName: "kube-api-access-k6tsr") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "kube-api-access-k6tsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.109722 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.110265 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "da2c16cb-5691-4679-8591-fdfdcee443e7" (UID: "da2c16cb-5691-4679-8591-fdfdcee443e7"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204827 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204863 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204874 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/da2c16cb-5691-4679-8591-fdfdcee443e7-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204885 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204897 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/da2c16cb-5691-4679-8591-fdfdcee443e7-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204908 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/da2c16cb-5691-4679-8591-fdfdcee443e7-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.204920 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6tsr\" (UniqueName: \"kubernetes.io/projected/da2c16cb-5691-4679-8591-fdfdcee443e7-kube-api-access-k6tsr\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.623106 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-bundle-2-build" event={"ID":"da2c16cb-5691-4679-8591-fdfdcee443e7","Type":"ContainerDied","Data":"36ddba62092a70f310bf7fe3397b72578834334d9af2792a2e445ddeaa5050b9"} Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.623167 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36ddba62092a70f310bf7fe3397b72578834334d9af2792a2e445ddeaa5050b9" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.623125 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-bundle-2-build" Mar 15 00:32:04 crc kubenswrapper[4861]: I0315 00:32:04.934790 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.017719 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxg85\" (UniqueName: \"kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85\") pod \"67bf6356-73f3-4534-84a0-6f5ea23d7c77\" (UID: \"67bf6356-73f3-4534-84a0-6f5ea23d7c77\") " Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.024921 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85" (OuterVolumeSpecName: "kube-api-access-rxg85") pod "67bf6356-73f3-4534-84a0-6f5ea23d7c77" (UID: "67bf6356-73f3-4534-84a0-6f5ea23d7c77"). InnerVolumeSpecName "kube-api-access-rxg85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.119976 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxg85\" (UniqueName: \"kubernetes.io/projected/67bf6356-73f3-4534-84a0-6f5ea23d7c77-kube-api-access-rxg85\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.633399 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" event={"ID":"67bf6356-73f3-4534-84a0-6f5ea23d7c77","Type":"ContainerDied","Data":"dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4"} Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.633879 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfb5b2da7a3a54afb521d698f4936d3914a39a9ec4d3a683dbf28f0d78a733a4" Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.633495 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558912-qwsn4" Mar 15 00:32:05 crc kubenswrapper[4861]: I0315 00:32:05.997956 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558906-slfcz"] Mar 15 00:32:06 crc kubenswrapper[4861]: I0315 00:32:06.003957 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558906-slfcz"] Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.419230 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd1741e-a90b-4025-ac04-e1da8c46ca7f" path="/var/lib/kubelet/pods/fdd1741e-a90b-4025-ac04-e1da8c46ca7f/volumes" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814296 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Mar 15 00:32:07 crc kubenswrapper[4861]: E0315 00:32:07.814593 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="git-clone" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814611 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="git-clone" Mar 15 00:32:07 crc kubenswrapper[4861]: E0315 00:32:07.814626 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="docker-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814633 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="docker-build" Mar 15 00:32:07 crc kubenswrapper[4861]: E0315 00:32:07.814647 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="manage-dockerfile" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814655 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="manage-dockerfile" Mar 15 00:32:07 crc kubenswrapper[4861]: E0315 00:32:07.814673 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67bf6356-73f3-4534-84a0-6f5ea23d7c77" containerName="oc" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814681 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="67bf6356-73f3-4534-84a0-6f5ea23d7c77" containerName="oc" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.814998 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="67bf6356-73f3-4534-84a0-6f5ea23d7c77" containerName="oc" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.815022 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="da2c16cb-5691-4679-8591-fdfdcee443e7" containerName="docker-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.815807 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.818944 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-sys-config" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.818944 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.818946 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-global-ca" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.819715 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-1-ca" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.829709 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.967728 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts7s4\" (UniqueName: \"kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968140 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968186 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968246 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968294 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968363 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968428 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968474 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968626 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968715 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968856 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:07 crc kubenswrapper[4861]: I0315 00:32:07.968921 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069789 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069858 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069893 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069924 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069965 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts7s4\" (UniqueName: \"kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.069986 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070007 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070033 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070054 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070082 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070110 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070132 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070659 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070952 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.070973 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.071001 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.071747 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.071918 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.072686 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.073017 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.073326 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.084418 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.084440 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.097977 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts7s4\" (UniqueName: \"kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4\") pod \"smart-gateway-operator-bundle-1-build\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.142518 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.377061 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Mar 15 00:32:08 crc kubenswrapper[4861]: I0315 00:32:08.655343 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"8c74437a-0af2-43f1-9d53-9cbe0263a34c","Type":"ContainerStarted","Data":"7f8f3a411d561eef8593476339034d26df91a65174c11a0d3863f24eb8b4bd35"} Mar 15 00:32:09 crc kubenswrapper[4861]: I0315 00:32:09.668030 4861 generic.go:334] "Generic (PLEG): container finished" podID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerID="ee7849619683fa8b50258e9053abc551cfc871a5c8b83b1374a8f0e39a22dc34" exitCode=0 Mar 15 00:32:09 crc kubenswrapper[4861]: I0315 00:32:09.668131 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"8c74437a-0af2-43f1-9d53-9cbe0263a34c","Type":"ContainerDied","Data":"ee7849619683fa8b50258e9053abc551cfc871a5c8b83b1374a8f0e39a22dc34"} Mar 15 00:32:10 crc kubenswrapper[4861]: I0315 00:32:10.680634 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_8c74437a-0af2-43f1-9d53-9cbe0263a34c/docker-build/0.log" Mar 15 00:32:10 crc kubenswrapper[4861]: I0315 00:32:10.681838 4861 generic.go:334] "Generic (PLEG): container finished" podID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerID="38f3b796ab2d462f3ad9f920452defd44ed7fded5c839d3e5745e9cfdf6784a8" exitCode=1 Mar 15 00:32:10 crc kubenswrapper[4861]: I0315 00:32:10.681903 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"8c74437a-0af2-43f1-9d53-9cbe0263a34c","Type":"ContainerDied","Data":"38f3b796ab2d462f3ad9f920452defd44ed7fded5c839d3e5745e9cfdf6784a8"} Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.928197 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_8c74437a-0af2-43f1-9d53-9cbe0263a34c/docker-build/0.log" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.928993 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971441 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971521 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971680 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971734 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971788 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971837 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.971917 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972009 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972071 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972134 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972189 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972239 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts7s4\" (UniqueName: \"kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4\") pod \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\" (UID: \"8c74437a-0af2-43f1-9d53-9cbe0263a34c\") " Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972492 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972525 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972586 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972616 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972825 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972855 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8c74437a-0af2-43f1-9d53-9cbe0263a34c-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972870 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.972883 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.973007 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.973528 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.973732 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.974664 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.975342 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.981669 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.981753 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4" (OuterVolumeSpecName: "kube-api-access-ts7s4") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "kube-api-access-ts7s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:32:11 crc kubenswrapper[4861]: I0315 00:32:11.987357 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "8c74437a-0af2-43f1-9d53-9cbe0263a34c" (UID: "8c74437a-0af2-43f1-9d53-9cbe0263a34c"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073670 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073719 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073738 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts7s4\" (UniqueName: \"kubernetes.io/projected/8c74437a-0af2-43f1-9d53-9cbe0263a34c-kube-api-access-ts7s4\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073755 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073773 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073790 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/8c74437a-0af2-43f1-9d53-9cbe0263a34c-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073809 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8c74437a-0af2-43f1-9d53-9cbe0263a34c-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.073827 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8c74437a-0af2-43f1-9d53-9cbe0263a34c-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.701800 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-1-build_8c74437a-0af2-43f1-9d53-9cbe0263a34c/docker-build/0.log" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.702382 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-1-build" event={"ID":"8c74437a-0af2-43f1-9d53-9cbe0263a34c","Type":"ContainerDied","Data":"7f8f3a411d561eef8593476339034d26df91a65174c11a0d3863f24eb8b4bd35"} Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.702427 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f8f3a411d561eef8593476339034d26df91a65174c11a0d3863f24eb8b4bd35" Mar 15 00:32:12 crc kubenswrapper[4861]: I0315 00:32:12.702544 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-1-build" Mar 15 00:32:18 crc kubenswrapper[4861]: I0315 00:32:18.268921 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Mar 15 00:32:18 crc kubenswrapper[4861]: I0315 00:32:18.279442 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-1-build"] Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.427984 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" path="/var/lib/kubelet/pods/8c74437a-0af2-43f1-9d53-9cbe0263a34c/volumes" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.943419 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Mar 15 00:32:19 crc kubenswrapper[4861]: E0315 00:32:19.943925 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerName="docker-build" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.943960 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerName="docker-build" Mar 15 00:32:19 crc kubenswrapper[4861]: E0315 00:32:19.944012 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerName="manage-dockerfile" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.944029 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerName="manage-dockerfile" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.944253 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c74437a-0af2-43f1-9d53-9cbe0263a34c" containerName="docker-build" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.946108 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.952206 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-sys-config" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.952722 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.952999 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-ca" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.953250 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-bundle-2-global-ca" Mar 15 00:32:19 crc kubenswrapper[4861]: I0315 00:32:19.974365 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011018 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011101 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011186 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011224 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011409 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011477 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011512 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011639 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011671 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011713 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011740 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdmj5\" (UniqueName: \"kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.011816 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112387 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112489 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112539 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112656 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112705 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112765 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112806 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdmj5\" (UniqueName: \"kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112863 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112939 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.112999 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.113079 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.113139 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114030 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114377 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114462 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114476 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114691 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.114839 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.115058 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.115635 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.115889 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.125039 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.126787 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.139148 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdmj5\" (UniqueName: \"kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5\") pod \"smart-gateway-operator-bundle-2-build\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.274447 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:20 crc kubenswrapper[4861]: I0315 00:32:20.821369 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-bundle-2-build"] Mar 15 00:32:21 crc kubenswrapper[4861]: I0315 00:32:21.802668 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerStarted","Data":"5215f1793aadadc52d712d927e9eac04eae0f6ff5f353425fba3b47adb461d55"} Mar 15 00:32:21 crc kubenswrapper[4861]: I0315 00:32:21.802753 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerStarted","Data":"ad027ca122e17a1a8ac104dbdbd944f829195670ece1bcbc9afe3fbd7aba26bb"} Mar 15 00:32:22 crc kubenswrapper[4861]: I0315 00:32:22.817924 4861 generic.go:334] "Generic (PLEG): container finished" podID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerID="5215f1793aadadc52d712d927e9eac04eae0f6ff5f353425fba3b47adb461d55" exitCode=0 Mar 15 00:32:22 crc kubenswrapper[4861]: I0315 00:32:22.818535 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerDied","Data":"5215f1793aadadc52d712d927e9eac04eae0f6ff5f353425fba3b47adb461d55"} Mar 15 00:32:23 crc kubenswrapper[4861]: I0315 00:32:23.833483 4861 generic.go:334] "Generic (PLEG): container finished" podID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerID="5cfa93d0b3b9b9d4d2929efb9d3db752e44f0d08dbf49b1461f703688b88ae13" exitCode=0 Mar 15 00:32:23 crc kubenswrapper[4861]: I0315 00:32:23.833648 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerDied","Data":"5cfa93d0b3b9b9d4d2929efb9d3db752e44f0d08dbf49b1461f703688b88ae13"} Mar 15 00:32:23 crc kubenswrapper[4861]: I0315 00:32:23.909045 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-bundle-2-build_90d2e0cc-5581-42a3-a50b-0c2319e79c52/manage-dockerfile/0.log" Mar 15 00:32:24 crc kubenswrapper[4861]: I0315 00:32:24.844755 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerStarted","Data":"0ac284e85c5dae212643b5aae4d3fe5e25463b227b162aa62af7b686c45a1d98"} Mar 15 00:32:24 crc kubenswrapper[4861]: I0315 00:32:24.877734 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-bundle-2-build" podStartSLOduration=5.877677411 podStartE2EDuration="5.877677411s" podCreationTimestamp="2026-03-15 00:32:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:32:24.869029657 +0000 UTC m=+1538.181227220" watchObservedRunningTime="2026-03-15 00:32:24.877677411 +0000 UTC m=+1538.189874984" Mar 15 00:32:27 crc kubenswrapper[4861]: I0315 00:32:27.876366 4861 generic.go:334] "Generic (PLEG): container finished" podID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerID="0ac284e85c5dae212643b5aae4d3fe5e25463b227b162aa62af7b686c45a1d98" exitCode=0 Mar 15 00:32:27 crc kubenswrapper[4861]: I0315 00:32:27.876426 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerDied","Data":"0ac284e85c5dae212643b5aae4d3fe5e25463b227b162aa62af7b686c45a1d98"} Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.244432 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373172 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373243 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373275 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373308 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373329 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373355 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373453 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373480 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373531 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdmj5\" (UniqueName: \"kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373575 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373598 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373700 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.373766 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run\") pod \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\" (UID: \"90d2e0cc-5581-42a3-a50b-0c2319e79c52\") " Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.374317 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.374398 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.374321 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.374926 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.375218 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.375311 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.375859 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.377850 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.381007 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.382149 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5" (OuterVolumeSpecName: "kube-api-access-xdmj5") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "kube-api-access-xdmj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.382405 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.382474 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "90d2e0cc-5581-42a3-a50b-0c2319e79c52" (UID: "90d2e0cc-5581-42a3-a50b-0c2319e79c52"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475751 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475796 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475810 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475823 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/90d2e0cc-5581-42a3-a50b-0c2319e79c52-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475838 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475852 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475865 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475878 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdmj5\" (UniqueName: \"kubernetes.io/projected/90d2e0cc-5581-42a3-a50b-0c2319e79c52-kube-api-access-xdmj5\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475890 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90d2e0cc-5581-42a3-a50b-0c2319e79c52-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475901 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/90d2e0cc-5581-42a3-a50b-0c2319e79c52-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.475912 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/90d2e0cc-5581-42a3-a50b-0c2319e79c52-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.898420 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-bundle-2-build" event={"ID":"90d2e0cc-5581-42a3-a50b-0c2319e79c52","Type":"ContainerDied","Data":"ad027ca122e17a1a8ac104dbdbd944f829195670ece1bcbc9afe3fbd7aba26bb"} Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.898482 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad027ca122e17a1a8ac104dbdbd944f829195670ece1bcbc9afe3fbd7aba26bb" Mar 15 00:32:29 crc kubenswrapper[4861]: I0315 00:32:29.898603 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-bundle-2-build" Mar 15 00:32:32 crc kubenswrapper[4861]: I0315 00:32:32.282607 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:32:32 crc kubenswrapper[4861]: I0315 00:32:32.282719 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.084064 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 15 00:32:47 crc kubenswrapper[4861]: E0315 00:32:47.084966 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="manage-dockerfile" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.084984 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="manage-dockerfile" Mar 15 00:32:47 crc kubenswrapper[4861]: E0315 00:32:47.084997 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="git-clone" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.085005 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="git-clone" Mar 15 00:32:47 crc kubenswrapper[4861]: E0315 00:32:47.085020 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="docker-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.085027 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="docker-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.085142 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d2e0cc-5581-42a3-a50b-0c2319e79c52" containerName="docker-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.085959 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.087815 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.088080 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.088114 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.088142 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.088144 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.100545 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.252372 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.252422 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.252448 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.252472 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vcs7\" (UniqueName: \"kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.252500 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253021 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253220 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253279 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253344 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253412 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253458 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253528 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.253603 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.354989 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355076 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355120 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355159 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355188 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355221 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355252 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355314 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355340 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355366 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355363 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355396 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vcs7\" (UniqueName: \"kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355540 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355665 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355773 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.355868 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.356130 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.356133 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.356543 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.359032 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.359222 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.359256 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-pgvms" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.359689 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.359867 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.366860 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.367277 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.368133 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.377893 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.377890 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.382084 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.392470 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vcs7\" (UniqueName: \"kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7\") pod \"service-telemetry-framework-index-1-build\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.449333 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:32:47 crc kubenswrapper[4861]: I0315 00:32:47.752525 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Mar 15 00:32:48 crc kubenswrapper[4861]: I0315 00:32:48.093016 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerStarted","Data":"7f9e2b2bc1e5b22829a0b3589120791c9a00776ec2a5c5a8257b6f7e129ba50a"} Mar 15 00:32:48 crc kubenswrapper[4861]: I0315 00:32:48.093741 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerStarted","Data":"f407084ce229ac8e60745fe97edc7cd89b1c30fcaa2b8b8e17f11fcecccd889f"} Mar 15 00:32:48 crc kubenswrapper[4861]: E0315 00:32:48.225811 4861 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:56744->38.102.83.174:45953: write tcp 38.102.83.174:56744->38.102.83.174:45953: write: broken pipe Mar 15 00:32:48 crc kubenswrapper[4861]: I0315 00:32:48.680933 4861 scope.go:117] "RemoveContainer" containerID="3ec4a821e25757a09d89c7e435091981f55950e1ca0a5dbcddaee3b19deabbc3" Mar 15 00:32:49 crc kubenswrapper[4861]: I0315 00:32:49.110868 4861 generic.go:334] "Generic (PLEG): container finished" podID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerID="7f9e2b2bc1e5b22829a0b3589120791c9a00776ec2a5c5a8257b6f7e129ba50a" exitCode=0 Mar 15 00:32:49 crc kubenswrapper[4861]: I0315 00:32:49.110966 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerDied","Data":"7f9e2b2bc1e5b22829a0b3589120791c9a00776ec2a5c5a8257b6f7e129ba50a"} Mar 15 00:32:50 crc kubenswrapper[4861]: I0315 00:32:50.125671 4861 generic.go:334] "Generic (PLEG): container finished" podID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerID="2e990e72031cde6c50ef6d6a5eaeb33ae3e725c1cefe9e81181b51927154eb2b" exitCode=0 Mar 15 00:32:50 crc kubenswrapper[4861]: I0315 00:32:50.125870 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerDied","Data":"2e990e72031cde6c50ef6d6a5eaeb33ae3e725c1cefe9e81181b51927154eb2b"} Mar 15 00:32:50 crc kubenswrapper[4861]: I0315 00:32:50.185864 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_20b4ece0-e7dc-4b90-9733-38cef63523b0/manage-dockerfile/0.log" Mar 15 00:32:51 crc kubenswrapper[4861]: I0315 00:32:51.139313 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerStarted","Data":"b2803893a88e2a87cf706f50b6d358b5eab329676396a38013a1eb87f320d944"} Mar 15 00:32:51 crc kubenswrapper[4861]: I0315 00:32:51.187022 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-framework-index-1-build" podStartSLOduration=4.186977296 podStartE2EDuration="4.186977296s" podCreationTimestamp="2026-03-15 00:32:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:32:51.179592367 +0000 UTC m=+1564.491789930" watchObservedRunningTime="2026-03-15 00:32:51.186977296 +0000 UTC m=+1564.499174869" Mar 15 00:33:02 crc kubenswrapper[4861]: I0315 00:33:02.282271 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:33:02 crc kubenswrapper[4861]: I0315 00:33:02.283322 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:33:19 crc kubenswrapper[4861]: I0315 00:33:19.377783 4861 generic.go:334] "Generic (PLEG): container finished" podID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerID="b2803893a88e2a87cf706f50b6d358b5eab329676396a38013a1eb87f320d944" exitCode=0 Mar 15 00:33:19 crc kubenswrapper[4861]: I0315 00:33:19.378062 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerDied","Data":"b2803893a88e2a87cf706f50b6d358b5eab329676396a38013a1eb87f320d944"} Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.751598 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.919936 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920483 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920532 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920613 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920725 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920778 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920881 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920941 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920997 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vcs7\" (UniqueName: \"kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.920994 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.921062 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.921074 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.921147 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.921237 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.921321 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles\") pod \"20b4ece0-e7dc-4b90-9733-38cef63523b0\" (UID: \"20b4ece0-e7dc-4b90-9733-38cef63523b0\") " Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.922344 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.922393 4861 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildcachedir\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.922438 4861 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/20b4ece0-e7dc-4b90-9733-38cef63523b0-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.923607 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.923648 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.923762 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.923880 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.931790 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push" (OuterVolumeSpecName: "builder-dockercfg-pgvms-push") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "builder-dockercfg-pgvms-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.932092 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull" (OuterVolumeSpecName: "builder-dockercfg-pgvms-pull") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "builder-dockercfg-pgvms-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.932378 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7" (OuterVolumeSpecName: "kube-api-access-4vcs7") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "kube-api-access-4vcs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:33:20 crc kubenswrapper[4861]: I0315 00:33:20.932174 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.025927 4861 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-system-configs\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026116 4861 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026151 4861 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026235 4861 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026317 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-run\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026403 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-pull\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-pull\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026482 4861 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-pgvms-push\" (UniqueName: \"kubernetes.io/secret/20b4ece0-e7dc-4b90-9733-38cef63523b0-builder-dockercfg-pgvms-push\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026515 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vcs7\" (UniqueName: \"kubernetes.io/projected/20b4ece0-e7dc-4b90-9733-38cef63523b0-kube-api-access-4vcs7\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.026541 4861 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-buildworkdir\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.178070 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.230080 4861 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-build-blob-cache\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.411293 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.419198 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"20b4ece0-e7dc-4b90-9733-38cef63523b0","Type":"ContainerDied","Data":"f407084ce229ac8e60745fe97edc7cd89b1c30fcaa2b8b8e17f11fcecccd889f"} Mar 15 00:33:21 crc kubenswrapper[4861]: I0315 00:33:21.419253 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f407084ce229ac8e60745fe97edc7cd89b1c30fcaa2b8b8e17f11fcecccd889f" Mar 15 00:33:22 crc kubenswrapper[4861]: I0315 00:33:22.435935 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "20b4ece0-e7dc-4b90-9733-38cef63523b0" (UID: "20b4ece0-e7dc-4b90-9733-38cef63523b0"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:33:22 crc kubenswrapper[4861]: I0315 00:33:22.449404 4861 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/20b4ece0-e7dc-4b90-9733-38cef63523b0-container-storage-root\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.141809 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:23 crc kubenswrapper[4861]: E0315 00:33:23.142088 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="git-clone" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.142388 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="git-clone" Mar 15 00:33:23 crc kubenswrapper[4861]: E0315 00:33:23.142409 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="manage-dockerfile" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.142415 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="manage-dockerfile" Mar 15 00:33:23 crc kubenswrapper[4861]: E0315 00:33:23.142431 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="docker-build" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.142437 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="docker-build" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.142537 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b4ece0-e7dc-4b90-9733-38cef63523b0" containerName="docker-build" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.143206 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.146812 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-m7pjz" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.157067 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.262361 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bbzd\" (UniqueName: \"kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd\") pod \"infrawatch-operators-tmpk9\" (UID: \"35ed58d8-f313-48cb-9142-72630585313f\") " pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.364849 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bbzd\" (UniqueName: \"kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd\") pod \"infrawatch-operators-tmpk9\" (UID: \"35ed58d8-f313-48cb-9142-72630585313f\") " pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.398383 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bbzd\" (UniqueName: \"kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd\") pod \"infrawatch-operators-tmpk9\" (UID: \"35ed58d8-f313-48cb-9142-72630585313f\") " pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.469731 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.967550 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:23 crc kubenswrapper[4861]: I0315 00:33:23.973924 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:33:24 crc kubenswrapper[4861]: I0315 00:33:24.442210 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-tmpk9" event={"ID":"35ed58d8-f313-48cb-9142-72630585313f","Type":"ContainerStarted","Data":"2bb99fd9a203e2587afbe3d17ecfeac87c61226a05480e4ab308e60be35c704a"} Mar 15 00:33:25 crc kubenswrapper[4861]: I0315 00:33:25.733299 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.541490 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-7vpqf"] Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.542974 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.548100 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-7vpqf"] Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.720586 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85ln2\" (UniqueName: \"kubernetes.io/projected/57ab3e10-e849-4e18-838e-7a94432a1521-kube-api-access-85ln2\") pod \"infrawatch-operators-7vpqf\" (UID: \"57ab3e10-e849-4e18-838e-7a94432a1521\") " pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.821990 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85ln2\" (UniqueName: \"kubernetes.io/projected/57ab3e10-e849-4e18-838e-7a94432a1521-kube-api-access-85ln2\") pod \"infrawatch-operators-7vpqf\" (UID: \"57ab3e10-e849-4e18-838e-7a94432a1521\") " pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.847318 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85ln2\" (UniqueName: \"kubernetes.io/projected/57ab3e10-e849-4e18-838e-7a94432a1521-kube-api-access-85ln2\") pod \"infrawatch-operators-7vpqf\" (UID: \"57ab3e10-e849-4e18-838e-7a94432a1521\") " pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:26 crc kubenswrapper[4861]: I0315 00:33:26.870288 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.282211 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.283266 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.283353 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.284210 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.284347 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" gracePeriod=600 Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.521516 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" exitCode=0 Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.521611 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187"} Mar 15 00:33:32 crc kubenswrapper[4861]: I0315 00:33:32.521663 4861 scope.go:117] "RemoveContainer" containerID="f8e6e1f9c412f0bf04b94830207e70efe443e435345ba67dd9721478768a5d19" Mar 15 00:33:36 crc kubenswrapper[4861]: E0315 00:33:36.080309 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:33:36 crc kubenswrapper[4861]: I0315 00:33:36.557377 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:33:36 crc kubenswrapper[4861]: E0315 00:33:36.558511 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:33:36 crc kubenswrapper[4861]: I0315 00:33:36.679491 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-7vpqf"] Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.579160 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-7vpqf" event={"ID":"57ab3e10-e849-4e18-838e-7a94432a1521","Type":"ContainerStarted","Data":"0ba1b32a5239e614e9b32b7de8fc4d17a37b4f1c9c484cc2babce6bedb6d4c7e"} Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.579240 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-7vpqf" event={"ID":"57ab3e10-e849-4e18-838e-7a94432a1521","Type":"ContainerStarted","Data":"5e7e360f1cda89c7cd4fa36688b7c9e4048e4a4e3cd3b69df3e4fba62885cbd3"} Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.581777 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-tmpk9" event={"ID":"35ed58d8-f313-48cb-9142-72630585313f","Type":"ContainerStarted","Data":"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85"} Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.582176 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-tmpk9" podUID="35ed58d8-f313-48cb-9142-72630585313f" containerName="registry-server" containerID="cri-o://1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85" gracePeriod=2 Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.619871 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-7vpqf" podStartSLOduration=11.494979304 podStartE2EDuration="11.6198165s" podCreationTimestamp="2026-03-15 00:33:26 +0000 UTC" firstStartedPulling="2026-03-15 00:33:36.683699487 +0000 UTC m=+1609.995897030" lastFinishedPulling="2026-03-15 00:33:36.808536693 +0000 UTC m=+1610.120734226" observedRunningTime="2026-03-15 00:33:37.608997507 +0000 UTC m=+1610.921195110" watchObservedRunningTime="2026-03-15 00:33:37.6198165 +0000 UTC m=+1610.932014073" Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.639967 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-tmpk9" podStartSLOduration=2.06934372 podStartE2EDuration="14.639931303s" podCreationTimestamp="2026-03-15 00:33:23 +0000 UTC" firstStartedPulling="2026-03-15 00:33:23.973641134 +0000 UTC m=+1597.285838677" lastFinishedPulling="2026-03-15 00:33:36.544228687 +0000 UTC m=+1609.856426260" observedRunningTime="2026-03-15 00:33:37.633280373 +0000 UTC m=+1610.945477916" watchObservedRunningTime="2026-03-15 00:33:37.639931303 +0000 UTC m=+1610.952128846" Mar 15 00:33:37 crc kubenswrapper[4861]: I0315 00:33:37.993020 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.056073 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bbzd\" (UniqueName: \"kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd\") pod \"35ed58d8-f313-48cb-9142-72630585313f\" (UID: \"35ed58d8-f313-48cb-9142-72630585313f\") " Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.062710 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd" (OuterVolumeSpecName: "kube-api-access-5bbzd") pod "35ed58d8-f313-48cb-9142-72630585313f" (UID: "35ed58d8-f313-48cb-9142-72630585313f"). InnerVolumeSpecName "kube-api-access-5bbzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.157882 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bbzd\" (UniqueName: \"kubernetes.io/projected/35ed58d8-f313-48cb-9142-72630585313f-kube-api-access-5bbzd\") on node \"crc\" DevicePath \"\"" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.591498 4861 generic.go:334] "Generic (PLEG): container finished" podID="35ed58d8-f313-48cb-9142-72630585313f" containerID="1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85" exitCode=0 Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.591581 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-tmpk9" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.591624 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-tmpk9" event={"ID":"35ed58d8-f313-48cb-9142-72630585313f","Type":"ContainerDied","Data":"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85"} Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.591705 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-tmpk9" event={"ID":"35ed58d8-f313-48cb-9142-72630585313f","Type":"ContainerDied","Data":"2bb99fd9a203e2587afbe3d17ecfeac87c61226a05480e4ab308e60be35c704a"} Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.591741 4861 scope.go:117] "RemoveContainer" containerID="1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.642663 4861 scope.go:117] "RemoveContainer" containerID="1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85" Mar 15 00:33:38 crc kubenswrapper[4861]: E0315 00:33:38.644098 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85\": container with ID starting with 1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85 not found: ID does not exist" containerID="1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.644135 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85"} err="failed to get container status \"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85\": rpc error: code = NotFound desc = could not find container \"1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85\": container with ID starting with 1c7dd52abbc365e56788af05932371f26cd958495d153a6279255f3f6fb25b85 not found: ID does not exist" Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.657163 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:38 crc kubenswrapper[4861]: I0315 00:33:38.663357 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-tmpk9"] Mar 15 00:33:39 crc kubenswrapper[4861]: I0315 00:33:39.425231 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ed58d8-f313-48cb-9142-72630585313f" path="/var/lib/kubelet/pods/35ed58d8-f313-48cb-9142-72630585313f/volumes" Mar 15 00:33:46 crc kubenswrapper[4861]: I0315 00:33:46.871047 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:46 crc kubenswrapper[4861]: I0315 00:33:46.871746 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:46 crc kubenswrapper[4861]: I0315 00:33:46.917647 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:33:47 crc kubenswrapper[4861]: I0315 00:33:47.420511 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:33:47 crc kubenswrapper[4861]: E0315 00:33:47.423189 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:33:47 crc kubenswrapper[4861]: I0315 00:33:47.716307 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-7vpqf" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.161286 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558914-69rrq"] Mar 15 00:34:00 crc kubenswrapper[4861]: E0315 00:34:00.162136 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ed58d8-f313-48cb-9142-72630585313f" containerName="registry-server" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.162151 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ed58d8-f313-48cb-9142-72630585313f" containerName="registry-server" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.162299 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ed58d8-f313-48cb-9142-72630585313f" containerName="registry-server" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.162944 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.165517 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.165681 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.166210 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.173195 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558914-69rrq"] Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.207261 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvc4s\" (UniqueName: \"kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s\") pod \"auto-csr-approver-29558914-69rrq\" (UID: \"b0d01275-16e4-416d-bfba-6a950f5fcb51\") " pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.309316 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvc4s\" (UniqueName: \"kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s\") pod \"auto-csr-approver-29558914-69rrq\" (UID: \"b0d01275-16e4-416d-bfba-6a950f5fcb51\") " pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.335724 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvc4s\" (UniqueName: \"kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s\") pod \"auto-csr-approver-29558914-69rrq\" (UID: \"b0d01275-16e4-416d-bfba-6a950f5fcb51\") " pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.408668 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:34:00 crc kubenswrapper[4861]: E0315 00:34:00.408974 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.483298 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:00 crc kubenswrapper[4861]: W0315 00:34:00.948461 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0d01275_16e4_416d_bfba_6a950f5fcb51.slice/crio-132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87 WatchSource:0}: Error finding container 132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87: Status 404 returned error can't find the container with id 132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87 Mar 15 00:34:00 crc kubenswrapper[4861]: I0315 00:34:00.950462 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558914-69rrq"] Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.220990 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57"] Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.223275 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.230260 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57"] Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.327487 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.328059 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.328211 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wxmh\" (UniqueName: \"kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.430183 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.430268 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wxmh\" (UniqueName: \"kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.430438 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.431250 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.431257 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.466352 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wxmh\" (UniqueName: \"kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh\") pod \"500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.549104 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.778659 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57"] Mar 15 00:34:01 crc kubenswrapper[4861]: W0315 00:34:01.781975 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66ebd7f3_2570_4365_a3dd_0e15e636a0b0.slice/crio-09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04 WatchSource:0}: Error finding container 09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04: Status 404 returned error can't find the container with id 09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04 Mar 15 00:34:01 crc kubenswrapper[4861]: I0315 00:34:01.789816 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558914-69rrq" event={"ID":"b0d01275-16e4-416d-bfba-6a950f5fcb51","Type":"ContainerStarted","Data":"132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87"} Mar 15 00:34:02 crc kubenswrapper[4861]: I0315 00:34:02.000350 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6"] Mar 15 00:34:02 crc kubenswrapper[4861]: I0315 00:34:02.001770 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.013076 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6"] Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.042680 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mcfz\" (UniqueName: \"kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.042758 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.042791 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.144860 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.144996 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mcfz\" (UniqueName: \"kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.145056 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.145438 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.145633 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.164780 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mcfz\" (UniqueName: \"kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz\") pod \"372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.318705 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.799602 4861 generic.go:334] "Generic (PLEG): container finished" podID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerID="cbce64724fe488a4c0b8a745fda2305e398d5b370e56c47f2a9a25d49fbbbc95" exitCode=0 Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.799659 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" event={"ID":"66ebd7f3-2570-4365-a3dd-0e15e636a0b0","Type":"ContainerDied","Data":"cbce64724fe488a4c0b8a745fda2305e398d5b370e56c47f2a9a25d49fbbbc95"} Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:02.799695 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" event={"ID":"66ebd7f3-2570-4365-a3dd-0e15e636a0b0","Type":"ContainerStarted","Data":"09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04"} Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.385647 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6"] Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.813928 4861 generic.go:334] "Generic (PLEG): container finished" podID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerID="032c1e2a93d4e5b0d901b085c8e86fa091934e30ee1e0b7f8dd7dc007070b16d" exitCode=0 Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.814125 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" event={"ID":"66ebd7f3-2570-4365-a3dd-0e15e636a0b0","Type":"ContainerDied","Data":"032c1e2a93d4e5b0d901b085c8e86fa091934e30ee1e0b7f8dd7dc007070b16d"} Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.819165 4861 generic.go:334] "Generic (PLEG): container finished" podID="b0d01275-16e4-416d-bfba-6a950f5fcb51" containerID="29a66d0c13c295ed5b878cb51639d59c20028afe3237e7e37425415742411364" exitCode=0 Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.819300 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558914-69rrq" event={"ID":"b0d01275-16e4-416d-bfba-6a950f5fcb51","Type":"ContainerDied","Data":"29a66d0c13c295ed5b878cb51639d59c20028afe3237e7e37425415742411364"} Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.825418 4861 generic.go:334] "Generic (PLEG): container finished" podID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerID="7dd27a2a3fdb69980f4b6fc1a800dff1b761cb3a5f630f1d28daad92445b8c81" exitCode=0 Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.825471 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" event={"ID":"e859ceca-5a54-4eeb-80da-ebd7acd61d9f","Type":"ContainerDied","Data":"7dd27a2a3fdb69980f4b6fc1a800dff1b761cb3a5f630f1d28daad92445b8c81"} Mar 15 00:34:03 crc kubenswrapper[4861]: I0315 00:34:03.825502 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" event={"ID":"e859ceca-5a54-4eeb-80da-ebd7acd61d9f","Type":"ContainerStarted","Data":"520c961f6b4872f17a8dac28e74ab427ea33e46007580063fdce5c05a7595ba1"} Mar 15 00:34:04 crc kubenswrapper[4861]: I0315 00:34:04.846388 4861 generic.go:334] "Generic (PLEG): container finished" podID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerID="f0def675d9df7d36308cae65db44a3a1c9f8d81d5c9708d82affbb6b3e4cb7bb" exitCode=0 Mar 15 00:34:04 crc kubenswrapper[4861]: I0315 00:34:04.846516 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" event={"ID":"66ebd7f3-2570-4365-a3dd-0e15e636a0b0","Type":"ContainerDied","Data":"f0def675d9df7d36308cae65db44a3a1c9f8d81d5c9708d82affbb6b3e4cb7bb"} Mar 15 00:34:04 crc kubenswrapper[4861]: I0315 00:34:04.849820 4861 generic.go:334] "Generic (PLEG): container finished" podID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerID="3cd3b8912f19686f2c1e60344c25727869f10d9d4727cdb3bc0aa853333690cc" exitCode=0 Mar 15 00:34:04 crc kubenswrapper[4861]: I0315 00:34:04.849904 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" event={"ID":"e859ceca-5a54-4eeb-80da-ebd7acd61d9f","Type":"ContainerDied","Data":"3cd3b8912f19686f2c1e60344c25727869f10d9d4727cdb3bc0aa853333690cc"} Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.119000 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.202783 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvc4s\" (UniqueName: \"kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s\") pod \"b0d01275-16e4-416d-bfba-6a950f5fcb51\" (UID: \"b0d01275-16e4-416d-bfba-6a950f5fcb51\") " Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.213199 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s" (OuterVolumeSpecName: "kube-api-access-kvc4s") pod "b0d01275-16e4-416d-bfba-6a950f5fcb51" (UID: "b0d01275-16e4-416d-bfba-6a950f5fcb51"). InnerVolumeSpecName "kube-api-access-kvc4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.305655 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvc4s\" (UniqueName: \"kubernetes.io/projected/b0d01275-16e4-416d-bfba-6a950f5fcb51-kube-api-access-kvc4s\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.864609 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558914-69rrq" Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.864620 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558914-69rrq" event={"ID":"b0d01275-16e4-416d-bfba-6a950f5fcb51","Type":"ContainerDied","Data":"132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87"} Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.866501 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="132c0a47bdf8d0f85939c998474dbf8c8152f1b0da3eec936774e238e7b14b87" Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.874855 4861 generic.go:334] "Generic (PLEG): container finished" podID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerID="4dac08241dda8eac8536d8adba57f0efc874cf9695c0145a4219ed05630466b8" exitCode=0 Mar 15 00:34:05 crc kubenswrapper[4861]: I0315 00:34:05.875010 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" event={"ID":"e859ceca-5a54-4eeb-80da-ebd7acd61d9f","Type":"ContainerDied","Data":"4dac08241dda8eac8536d8adba57f0efc874cf9695c0145a4219ed05630466b8"} Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.190460 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558908-6l8x4"] Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.199645 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558908-6l8x4"] Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.216692 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.319921 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle\") pod \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.320041 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wxmh\" (UniqueName: \"kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh\") pod \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.320117 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util\") pod \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\" (UID: \"66ebd7f3-2570-4365-a3dd-0e15e636a0b0\") " Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.322116 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle" (OuterVolumeSpecName: "bundle") pod "66ebd7f3-2570-4365-a3dd-0e15e636a0b0" (UID: "66ebd7f3-2570-4365-a3dd-0e15e636a0b0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.325250 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh" (OuterVolumeSpecName: "kube-api-access-2wxmh") pod "66ebd7f3-2570-4365-a3dd-0e15e636a0b0" (UID: "66ebd7f3-2570-4365-a3dd-0e15e636a0b0"). InnerVolumeSpecName "kube-api-access-2wxmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.348470 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util" (OuterVolumeSpecName: "util") pod "66ebd7f3-2570-4365-a3dd-0e15e636a0b0" (UID: "66ebd7f3-2570-4365-a3dd-0e15e636a0b0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.421848 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.421907 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wxmh\" (UniqueName: \"kubernetes.io/projected/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-kube-api-access-2wxmh\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.421926 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/66ebd7f3-2570-4365-a3dd-0e15e636a0b0-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.885943 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.885957 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/500c4f010310dad14c569d8fa2124fef1cf701af50ed1128cec4daf65avlc57" event={"ID":"66ebd7f3-2570-4365-a3dd-0e15e636a0b0","Type":"ContainerDied","Data":"09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04"} Mar 15 00:34:06 crc kubenswrapper[4861]: I0315 00:34:06.886679 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09af608ee660ebf943755383c72202bb7737584ef5980581980a6e0671ccac04" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.153482 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.232966 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mcfz\" (UniqueName: \"kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz\") pod \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.233172 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle\") pod \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.233376 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util\") pod \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\" (UID: \"e859ceca-5a54-4eeb-80da-ebd7acd61d9f\") " Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.233769 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle" (OuterVolumeSpecName: "bundle") pod "e859ceca-5a54-4eeb-80da-ebd7acd61d9f" (UID: "e859ceca-5a54-4eeb-80da-ebd7acd61d9f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.234233 4861 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-bundle\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.237548 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz" (OuterVolumeSpecName: "kube-api-access-7mcfz") pod "e859ceca-5a54-4eeb-80da-ebd7acd61d9f" (UID: "e859ceca-5a54-4eeb-80da-ebd7acd61d9f"). InnerVolumeSpecName "kube-api-access-7mcfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.248578 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util" (OuterVolumeSpecName: "util") pod "e859ceca-5a54-4eeb-80da-ebd7acd61d9f" (UID: "e859ceca-5a54-4eeb-80da-ebd7acd61d9f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.336009 4861 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-util\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.336202 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mcfz\" (UniqueName: \"kubernetes.io/projected/e859ceca-5a54-4eeb-80da-ebd7acd61d9f-kube-api-access-7mcfz\") on node \"crc\" DevicePath \"\"" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.422425 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ef5f4c-e426-43d2-90f8-c21be80ece91" path="/var/lib/kubelet/pods/a1ef5f4c-e426-43d2-90f8-c21be80ece91/volumes" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.898265 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" event={"ID":"e859ceca-5a54-4eeb-80da-ebd7acd61d9f","Type":"ContainerDied","Data":"520c961f6b4872f17a8dac28e74ab427ea33e46007580063fdce5c05a7595ba1"} Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.898672 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="520c961f6b4872f17a8dac28e74ab427ea33e46007580063fdce5c05a7595ba1" Mar 15 00:34:07 crc kubenswrapper[4861]: I0315 00:34:07.898417 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/372e7d5daac88c2e9a91443a2f508c8c20ad57bc41b1606ec960d61c09plqq6" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.098438 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs"] Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099316 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="pull" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099339 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="pull" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099354 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="util" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099364 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="util" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099375 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099385 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099403 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="pull" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099412 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="pull" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099428 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099437 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099449 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0d01275-16e4-416d-bfba-6a950f5fcb51" containerName="oc" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099457 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0d01275-16e4-416d-bfba-6a950f5fcb51" containerName="oc" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.099471 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="util" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099497 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="util" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099665 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="66ebd7f3-2570-4365-a3dd-0e15e636a0b0" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099693 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="e859ceca-5a54-4eeb-80da-ebd7acd61d9f" containerName="extract" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.099706 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0d01275-16e4-416d-bfba-6a950f5fcb51" containerName="oc" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.100343 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.102795 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-2gq7r" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.115886 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs"] Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.210409 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/ee2860e1-04df-4782-9bf0-0baf205e595d-runner\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.210466 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2dbl\" (UniqueName: \"kubernetes.io/projected/ee2860e1-04df-4782-9bf0-0baf205e595d-kube-api-access-v2dbl\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.312243 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/ee2860e1-04df-4782-9bf0-0baf205e595d-runner\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.312310 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2dbl\" (UniqueName: \"kubernetes.io/projected/ee2860e1-04df-4782-9bf0-0baf205e595d-kube-api-access-v2dbl\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.312995 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/ee2860e1-04df-4782-9bf0-0baf205e595d-runner\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.340385 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2dbl\" (UniqueName: \"kubernetes.io/projected/ee2860e1-04df-4782-9bf0-0baf205e595d-kube-api-access-v2dbl\") pod \"service-telemetry-operator-86bfdfcc6-jwwzs\" (UID: \"ee2860e1-04df-4782-9bf0-0baf205e595d\") " pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.409789 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:34:12 crc kubenswrapper[4861]: E0315 00:34:12.410059 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.421364 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.923207 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs"] Mar 15 00:34:12 crc kubenswrapper[4861]: I0315 00:34:12.943386 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" event={"ID":"ee2860e1-04df-4782-9bf0-0baf205e595d","Type":"ContainerStarted","Data":"c451a0d13bffbeb3e7c97f44d793acc4b6aeb0e9614ef45bcaa2c7ce8b0eecb0"} Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.044421 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k"] Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.047470 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.053283 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-9m9p5" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.064010 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k"] Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.191521 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nlj\" (UniqueName: \"kubernetes.io/projected/1ea04940-819c-401c-842f-86b7b48ee8eb-kube-api-access-98nlj\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.191603 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ea04940-819c-401c-842f-86b7b48ee8eb-runner\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.293575 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98nlj\" (UniqueName: \"kubernetes.io/projected/1ea04940-819c-401c-842f-86b7b48ee8eb-kube-api-access-98nlj\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.293648 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ea04940-819c-401c-842f-86b7b48ee8eb-runner\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.294474 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/1ea04940-819c-401c-842f-86b7b48ee8eb-runner\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.321775 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98nlj\" (UniqueName: \"kubernetes.io/projected/1ea04940-819c-401c-842f-86b7b48ee8eb-kube-api-access-98nlj\") pod \"smart-gateway-operator-5fcfd8dffb-6sd8k\" (UID: \"1ea04940-819c-401c-842f-86b7b48ee8eb\") " pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.422257 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.892778 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k"] Mar 15 00:34:15 crc kubenswrapper[4861]: I0315 00:34:15.983916 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" event={"ID":"1ea04940-819c-401c-842f-86b7b48ee8eb","Type":"ContainerStarted","Data":"016f1e90f3a3d56fe4b1e3f0f38f1b6d4a607f8beee59da578c95eacca67162e"} Mar 15 00:34:26 crc kubenswrapper[4861]: I0315 00:34:26.408865 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:34:26 crc kubenswrapper[4861]: E0315 00:34:26.409536 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:34:30 crc kubenswrapper[4861]: E0315 00:34:30.752446 4861 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/service-telemetry-operator:latest" Mar 15 00:34:30 crc kubenswrapper[4861]: E0315 00:34:30.753179 4861 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/service-telemetry-operator:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:service-telemetry-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_WEBHOOK_SNMP_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/prometheus-webhook-snmp:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_IMAGE,Value:quay.io/prometheus/prometheus:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER_IMAGE,Value:quay.io/prometheus/alertmanager:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:service-telemetry-operator.v1.5.1773534697,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v2dbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod service-telemetry-operator-86bfdfcc6-jwwzs_service-telemetry(ee2860e1-04df-4782-9bf0-0baf205e595d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 15 00:34:30 crc kubenswrapper[4861]: E0315 00:34:30.754726 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" podUID="ee2860e1-04df-4782-9bf0-0baf205e595d" Mar 15 00:34:31 crc kubenswrapper[4861]: E0315 00:34:31.114492 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/service-telemetry-operator:latest\\\"\"" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" podUID="ee2860e1-04df-4782-9bf0-0baf205e595d" Mar 15 00:34:37 crc kubenswrapper[4861]: I0315 00:34:37.181391 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" event={"ID":"1ea04940-819c-401c-842f-86b7b48ee8eb","Type":"ContainerStarted","Data":"d8b07177367ef307382c4d4d919800012616e2f243b76b46577732575773d8aa"} Mar 15 00:34:37 crc kubenswrapper[4861]: I0315 00:34:37.218629 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-5fcfd8dffb-6sd8k" podStartSLOduration=1.510428863 podStartE2EDuration="22.218599559s" podCreationTimestamp="2026-03-15 00:34:15 +0000 UTC" firstStartedPulling="2026-03-15 00:34:15.904954221 +0000 UTC m=+1649.217151754" lastFinishedPulling="2026-03-15 00:34:36.613124907 +0000 UTC m=+1669.925322450" observedRunningTime="2026-03-15 00:34:37.210658274 +0000 UTC m=+1670.522855887" watchObservedRunningTime="2026-03-15 00:34:37.218599559 +0000 UTC m=+1670.530797102" Mar 15 00:34:40 crc kubenswrapper[4861]: I0315 00:34:40.408436 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:34:40 crc kubenswrapper[4861]: E0315 00:34:40.409019 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:34:47 crc kubenswrapper[4861]: I0315 00:34:47.272168 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" event={"ID":"ee2860e1-04df-4782-9bf0-0baf205e595d","Type":"ContainerStarted","Data":"9674d9512b19f75b06acb7f752b1b4e9f8fa41a540b870fe381d6f5aec127fdf"} Mar 15 00:34:47 crc kubenswrapper[4861]: I0315 00:34:47.297120 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-86bfdfcc6-jwwzs" podStartSLOduration=1.198213703 podStartE2EDuration="35.297098886s" podCreationTimestamp="2026-03-15 00:34:12 +0000 UTC" firstStartedPulling="2026-03-15 00:34:12.933170836 +0000 UTC m=+1646.245368389" lastFinishedPulling="2026-03-15 00:34:47.032056029 +0000 UTC m=+1680.344253572" observedRunningTime="2026-03-15 00:34:47.29062089 +0000 UTC m=+1680.602818433" watchObservedRunningTime="2026-03-15 00:34:47.297098886 +0000 UTC m=+1680.609296429" Mar 15 00:34:48 crc kubenswrapper[4861]: I0315 00:34:48.825682 4861 scope.go:117] "RemoveContainer" containerID="b5dd2c42d411ad4fc426dfd17121ea463207a4b2c8d151104992489dd91731a3" Mar 15 00:34:55 crc kubenswrapper[4861]: I0315 00:34:55.409880 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:34:55 crc kubenswrapper[4861]: E0315 00:34:55.410938 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:08 crc kubenswrapper[4861]: I0315 00:35:08.409283 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:35:08 crc kubenswrapper[4861]: E0315 00:35:08.410174 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.606926 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.608839 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.612449 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.612475 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-7fdxt" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.614860 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.615305 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.615607 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.615925 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.616183 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.626660 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696203 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696255 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696285 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdmp5\" (UniqueName: \"kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696301 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696321 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696339 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.696387 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797192 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797263 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797287 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797317 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdmp5\" (UniqueName: \"kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797338 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797356 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.797375 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.800087 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.803111 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.804976 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.805533 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.805688 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.806761 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.814378 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdmp5\" (UniqueName: \"kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5\") pod \"default-interconnect-68864d46cb-7487t\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:15 crc kubenswrapper[4861]: I0315 00:35:15.947174 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:35:16 crc kubenswrapper[4861]: I0315 00:35:16.468926 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:35:16 crc kubenswrapper[4861]: I0315 00:35:16.538605 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-7487t" event={"ID":"2ae02148-c9b7-4035-abc2-1f86adce42ec","Type":"ContainerStarted","Data":"b1cf3be3f80a2d911866f0296f5d6b5d6ed77bcfba2b9539a88049314b31575e"} Mar 15 00:35:19 crc kubenswrapper[4861]: I0315 00:35:19.409514 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:35:19 crc kubenswrapper[4861]: E0315 00:35:19.410140 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:21 crc kubenswrapper[4861]: I0315 00:35:21.604627 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-7487t" event={"ID":"2ae02148-c9b7-4035-abc2-1f86adce42ec","Type":"ContainerStarted","Data":"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6"} Mar 15 00:35:21 crc kubenswrapper[4861]: I0315 00:35:21.649214 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-7487t" podStartSLOduration=2.297503817 podStartE2EDuration="6.649177303s" podCreationTimestamp="2026-03-15 00:35:15 +0000 UTC" firstStartedPulling="2026-03-15 00:35:16.470650039 +0000 UTC m=+1709.782847602" lastFinishedPulling="2026-03-15 00:35:20.822323555 +0000 UTC m=+1714.134521088" observedRunningTime="2026-03-15 00:35:21.636368557 +0000 UTC m=+1714.948566140" watchObservedRunningTime="2026-03-15 00:35:21.649177303 +0000 UTC m=+1714.961374866" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.164859 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.168412 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.175152 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.175477 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.175657 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.175899 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-2" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.176059 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-1" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.176762 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.176873 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-bgplb" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.176992 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.177094 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.179314 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.199908 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292388 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292487 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb6br\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-kube-api-access-tb6br\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292523 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292591 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-tls-assets\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292631 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-web-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292653 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292681 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292707 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-75ae9852-7261-4488-a547-aaefae57ae9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-75ae9852-7261-4488-a547-aaefae57ae9a\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292748 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ef5a360a-3596-48be-8d28-ee75cfee0900-config-out\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292773 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292816 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.292846 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394613 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ef5a360a-3596-48be-8d28-ee75cfee0900-config-out\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394688 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394732 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394768 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394808 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394851 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb6br\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-kube-api-access-tb6br\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394876 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394898 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-tls-assets\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394934 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-web-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394955 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.394986 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.395012 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-75ae9852-7261-4488-a547-aaefae57ae9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-75ae9852-7261-4488-a547-aaefae57ae9a\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.397347 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-2\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: E0315 00:35:26.397474 4861 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 15 00:35:26 crc kubenswrapper[4861]: E0315 00:35:26.397545 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls podName:ef5a360a-3596-48be-8d28-ee75cfee0900 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:26.897508485 +0000 UTC m=+1720.209706018 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "ef5a360a-3596-48be-8d28-ee75cfee0900") : secret "default-prometheus-proxy-tls" not found Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.398714 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.398785 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-1\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.399481 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/ef5a360a-3596-48be-8d28-ee75cfee0900-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.401764 4861 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.401809 4861 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-75ae9852-7261-4488-a547-aaefae57ae9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-75ae9852-7261-4488-a547-aaefae57ae9a\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/278374575006b478b8b767bb1efab659a04611340902dd0b2f9789caa6f4e022/globalmount\"" pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.402331 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-tls-assets\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.402601 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ef5a360a-3596-48be-8d28-ee75cfee0900-config-out\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.402870 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-web-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.412089 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.418146 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-config\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.419061 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb6br\" (UniqueName: \"kubernetes.io/projected/ef5a360a-3596-48be-8d28-ee75cfee0900-kube-api-access-tb6br\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.438024 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-75ae9852-7261-4488-a547-aaefae57ae9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-75ae9852-7261-4488-a547-aaefae57ae9a\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: I0315 00:35:26.910749 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:26 crc kubenswrapper[4861]: E0315 00:35:26.910903 4861 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Mar 15 00:35:26 crc kubenswrapper[4861]: E0315 00:35:26.910979 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls podName:ef5a360a-3596-48be-8d28-ee75cfee0900 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:27.910960099 +0000 UTC m=+1721.223157632 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "ef5a360a-3596-48be-8d28-ee75cfee0900") : secret "default-prometheus-proxy-tls" not found Mar 15 00:35:27 crc kubenswrapper[4861]: I0315 00:35:27.923300 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:27 crc kubenswrapper[4861]: I0315 00:35:27.932253 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ef5a360a-3596-48be-8d28-ee75cfee0900-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"ef5a360a-3596-48be-8d28-ee75cfee0900\") " pod="service-telemetry/prometheus-default-0" Mar 15 00:35:28 crc kubenswrapper[4861]: I0315 00:35:28.049943 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Mar 15 00:35:28 crc kubenswrapper[4861]: I0315 00:35:28.404395 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Mar 15 00:35:28 crc kubenswrapper[4861]: I0315 00:35:28.674405 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerStarted","Data":"a7b1e405da3fdda9db4e9f560b03054b3f1a290a208254bd19d4ec153b42fbfc"} Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.410045 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:35:30 crc kubenswrapper[4861]: E0315 00:35:30.410764 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.748476 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.751988 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.782673 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.879828 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.879917 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wvkq\" (UniqueName: \"kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.880035 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.981779 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.982105 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wvkq\" (UniqueName: \"kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.982248 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.982284 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:30 crc kubenswrapper[4861]: I0315 00:35:30.982891 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.005997 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wvkq\" (UniqueName: \"kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq\") pod \"certified-operators-wshl8\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.088747 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.359933 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.728011 4861 generic.go:334] "Generic (PLEG): container finished" podID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerID="69544989e2c7bda5baae96d69b702ce0b66ab0ef831143619779d0ce7454e285" exitCode=0 Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.728185 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerDied","Data":"69544989e2c7bda5baae96d69b702ce0b66ab0ef831143619779d0ce7454e285"} Mar 15 00:35:31 crc kubenswrapper[4861]: I0315 00:35:31.728314 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerStarted","Data":"3082ffe6cff0e855d313e7169b1d4cf701c3c3d710698aea4cb963bcaca8269d"} Mar 15 00:35:32 crc kubenswrapper[4861]: I0315 00:35:32.739371 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerStarted","Data":"92813f4250324c3867f9072b2e866deb483250532c9ac08ed10931ac7bb9bf39"} Mar 15 00:35:32 crc kubenswrapper[4861]: I0315 00:35:32.741485 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerStarted","Data":"74c21026914f29f6281d9e2fe0ea7b385d878cec2fb0171e7f4bbc2916daba75"} Mar 15 00:35:33 crc kubenswrapper[4861]: I0315 00:35:33.754363 4861 generic.go:334] "Generic (PLEG): container finished" podID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerID="92813f4250324c3867f9072b2e866deb483250532c9ac08ed10931ac7bb9bf39" exitCode=0 Mar 15 00:35:33 crc kubenswrapper[4861]: I0315 00:35:33.754697 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerDied","Data":"92813f4250324c3867f9072b2e866deb483250532c9ac08ed10931ac7bb9bf39"} Mar 15 00:35:35 crc kubenswrapper[4861]: I0315 00:35:35.779702 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerStarted","Data":"546e4b0c4bb45333401ffa38208a7b710b5dc9f2a074b3af5a38fb443b879d7f"} Mar 15 00:35:35 crc kubenswrapper[4861]: I0315 00:35:35.821256 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wshl8" podStartSLOduration=2.939018235 podStartE2EDuration="5.821224718s" podCreationTimestamp="2026-03-15 00:35:30 +0000 UTC" firstStartedPulling="2026-03-15 00:35:31.73690785 +0000 UTC m=+1725.049105373" lastFinishedPulling="2026-03-15 00:35:34.619114283 +0000 UTC m=+1727.931311856" observedRunningTime="2026-03-15 00:35:35.81247182 +0000 UTC m=+1729.124669363" watchObservedRunningTime="2026-03-15 00:35:35.821224718 +0000 UTC m=+1729.133422291" Mar 15 00:35:36 crc kubenswrapper[4861]: I0315 00:35:36.833917 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-qs59c"] Mar 15 00:35:36 crc kubenswrapper[4861]: I0315 00:35:36.835485 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" Mar 15 00:35:36 crc kubenswrapper[4861]: I0315 00:35:36.851431 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-qs59c"] Mar 15 00:35:36 crc kubenswrapper[4861]: I0315 00:35:36.876608 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l88ww\" (UniqueName: \"kubernetes.io/projected/32144527-d295-4a45-ba18-ba6f391d60a9-kube-api-access-l88ww\") pod \"default-snmp-webhook-6856cfb745-qs59c\" (UID: \"32144527-d295-4a45-ba18-ba6f391d60a9\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" Mar 15 00:35:36 crc kubenswrapper[4861]: I0315 00:35:36.978577 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l88ww\" (UniqueName: \"kubernetes.io/projected/32144527-d295-4a45-ba18-ba6f391d60a9-kube-api-access-l88ww\") pod \"default-snmp-webhook-6856cfb745-qs59c\" (UID: \"32144527-d295-4a45-ba18-ba6f391d60a9\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" Mar 15 00:35:37 crc kubenswrapper[4861]: I0315 00:35:37.013015 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l88ww\" (UniqueName: \"kubernetes.io/projected/32144527-d295-4a45-ba18-ba6f391d60a9-kube-api-access-l88ww\") pod \"default-snmp-webhook-6856cfb745-qs59c\" (UID: \"32144527-d295-4a45-ba18-ba6f391d60a9\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" Mar 15 00:35:37 crc kubenswrapper[4861]: I0315 00:35:37.150818 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" Mar 15 00:35:37 crc kubenswrapper[4861]: I0315 00:35:37.398947 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-qs59c"] Mar 15 00:35:37 crc kubenswrapper[4861]: I0315 00:35:37.805344 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" event={"ID":"32144527-d295-4a45-ba18-ba6f391d60a9","Type":"ContainerStarted","Data":"0702b1c83ac069a04534f56d2c30d34fa1fd533fb449ca586627fc2b6a08a92c"} Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.153000 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.154803 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.157516 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-lk849" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.158091 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.158308 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.158448 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-cluster-tls-config" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.158597 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.159261 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.181534 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242005 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-web-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242083 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242144 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242172 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/931f3bda-5390-448a-b844-f198f756dcfa-config-out\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242206 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-config-volume\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242252 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242271 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzq2r\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-kube-api-access-nzq2r\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242400 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-tls-assets\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.242524 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344580 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-web-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344654 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344683 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/931f3bda-5390-448a-b844-f198f756dcfa-config-out\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344703 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344729 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-config-volume\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344759 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344777 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzq2r\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-kube-api-access-nzq2r\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344800 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-tls-assets\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.344826 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: E0315 00:35:40.345976 4861 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:40 crc kubenswrapper[4861]: E0315 00:35:40.346040 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls podName:931f3bda-5390-448a-b844-f198f756dcfa nodeName:}" failed. No retries permitted until 2026-03-15 00:35:40.846021995 +0000 UTC m=+1734.158219538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "931f3bda-5390-448a-b844-f198f756dcfa") : secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.353342 4861 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.354216 4861 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ac44325e94b3393f20cfecd5534d23953e41d60af6513d296c8113ec2087bf64/globalmount\"" pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.355482 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-config-volume\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.356230 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.356521 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/931f3bda-5390-448a-b844-f198f756dcfa-config-out\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.357091 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-web-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.362227 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-tls-assets\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.369779 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzq2r\" (UniqueName: \"kubernetes.io/projected/931f3bda-5390-448a-b844-f198f756dcfa-kube-api-access-nzq2r\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.374453 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.386047 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b3361ba-0651-44fc-8a9a-f10ccdcb9c8a\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: I0315 00:35:40.854283 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:40 crc kubenswrapper[4861]: E0315 00:35:40.854521 4861 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:40 crc kubenswrapper[4861]: E0315 00:35:40.854630 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls podName:931f3bda-5390-448a-b844-f198f756dcfa nodeName:}" failed. No retries permitted until 2026-03-15 00:35:41.854608418 +0000 UTC m=+1735.166805951 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "931f3bda-5390-448a-b844-f198f756dcfa") : secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.088886 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.089076 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.152201 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.856839 4861 generic.go:334] "Generic (PLEG): container finished" podID="ef5a360a-3596-48be-8d28-ee75cfee0900" containerID="74c21026914f29f6281d9e2fe0ea7b385d878cec2fb0171e7f4bbc2916daba75" exitCode=0 Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.856912 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerDied","Data":"74c21026914f29f6281d9e2fe0ea7b385d878cec2fb0171e7f4bbc2916daba75"} Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.869596 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:41 crc kubenswrapper[4861]: E0315 00:35:41.870276 4861 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:41 crc kubenswrapper[4861]: E0315 00:35:41.870327 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls podName:931f3bda-5390-448a-b844-f198f756dcfa nodeName:}" failed. No retries permitted until 2026-03-15 00:35:43.870310531 +0000 UTC m=+1737.182508064 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "931f3bda-5390-448a-b844-f198f756dcfa") : secret "default-alertmanager-proxy-tls" not found Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.919521 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:41 crc kubenswrapper[4861]: I0315 00:35:41.972458 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:43 crc kubenswrapper[4861]: I0315 00:35:43.408670 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:35:43 crc kubenswrapper[4861]: E0315 00:35:43.409108 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:43 crc kubenswrapper[4861]: I0315 00:35:43.875211 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wshl8" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="registry-server" containerID="cri-o://546e4b0c4bb45333401ffa38208a7b710b5dc9f2a074b3af5a38fb443b879d7f" gracePeriod=2 Mar 15 00:35:43 crc kubenswrapper[4861]: I0315 00:35:43.899019 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:43 crc kubenswrapper[4861]: I0315 00:35:43.907168 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/931f3bda-5390-448a-b844-f198f756dcfa-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"931f3bda-5390-448a-b844-f198f756dcfa\") " pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:44 crc kubenswrapper[4861]: I0315 00:35:44.097057 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Mar 15 00:35:44 crc kubenswrapper[4861]: I0315 00:35:44.884540 4861 generic.go:334] "Generic (PLEG): container finished" podID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerID="546e4b0c4bb45333401ffa38208a7b710b5dc9f2a074b3af5a38fb443b879d7f" exitCode=0 Mar 15 00:35:44 crc kubenswrapper[4861]: I0315 00:35:44.884606 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerDied","Data":"546e4b0c4bb45333401ffa38208a7b710b5dc9f2a074b3af5a38fb443b879d7f"} Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.090882 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.189184 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wvkq\" (UniqueName: \"kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq\") pod \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.189290 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content\") pod \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.189363 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities\") pod \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\" (UID: \"da1028fa-8876-4981-a6e0-f2f2e5e7d18c\") " Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.190874 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities" (OuterVolumeSpecName: "utilities") pod "da1028fa-8876-4981-a6e0-f2f2e5e7d18c" (UID: "da1028fa-8876-4981-a6e0-f2f2e5e7d18c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.205238 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq" (OuterVolumeSpecName: "kube-api-access-8wvkq") pod "da1028fa-8876-4981-a6e0-f2f2e5e7d18c" (UID: "da1028fa-8876-4981-a6e0-f2f2e5e7d18c"). InnerVolumeSpecName "kube-api-access-8wvkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.260680 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da1028fa-8876-4981-a6e0-f2f2e5e7d18c" (UID: "da1028fa-8876-4981-a6e0-f2f2e5e7d18c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.291431 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wvkq\" (UniqueName: \"kubernetes.io/projected/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-kube-api-access-8wvkq\") on node \"crc\" DevicePath \"\"" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.291462 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.291473 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da1028fa-8876-4981-a6e0-f2f2e5e7d18c-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.778204 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.919457 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" event={"ID":"32144527-d295-4a45-ba18-ba6f391d60a9","Type":"ContainerStarted","Data":"36d3e0764d7419fc206c9a6456efca6515abcb587f2ef4d50ef65382902d32c4"} Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.924687 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerStarted","Data":"98927ace59478baf02ca82582d324bb118519ccf382f7a9d0c184eee615a8adb"} Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.928086 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wshl8" event={"ID":"da1028fa-8876-4981-a6e0-f2f2e5e7d18c","Type":"ContainerDied","Data":"3082ffe6cff0e855d313e7169b1d4cf701c3c3d710698aea4cb963bcaca8269d"} Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.928131 4861 scope.go:117] "RemoveContainer" containerID="546e4b0c4bb45333401ffa38208a7b710b5dc9f2a074b3af5a38fb443b879d7f" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.928154 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wshl8" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.945567 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6856cfb745-qs59c" podStartSLOduration=2.057081553 podStartE2EDuration="12.945532191s" podCreationTimestamp="2026-03-15 00:35:36 +0000 UTC" firstStartedPulling="2026-03-15 00:35:37.419410111 +0000 UTC m=+1730.731607634" lastFinishedPulling="2026-03-15 00:35:48.307860739 +0000 UTC m=+1741.620058272" observedRunningTime="2026-03-15 00:35:48.941806259 +0000 UTC m=+1742.254003792" watchObservedRunningTime="2026-03-15 00:35:48.945532191 +0000 UTC m=+1742.257729724" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.959737 4861 scope.go:117] "RemoveContainer" containerID="92813f4250324c3867f9072b2e866deb483250532c9ac08ed10931ac7bb9bf39" Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.991176 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:48 crc kubenswrapper[4861]: I0315 00:35:48.992910 4861 scope.go:117] "RemoveContainer" containerID="69544989e2c7bda5baae96d69b702ce0b66ab0ef831143619779d0ce7454e285" Mar 15 00:35:49 crc kubenswrapper[4861]: I0315 00:35:49.004461 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wshl8"] Mar 15 00:35:49 crc kubenswrapper[4861]: I0315 00:35:49.421300 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" path="/var/lib/kubelet/pods/da1028fa-8876-4981-a6e0-f2f2e5e7d18c/volumes" Mar 15 00:35:50 crc kubenswrapper[4861]: I0315 00:35:50.951782 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerStarted","Data":"ba0d4c0656927866a201deb6de1d8e1836236cd55b6d541b0f0d2b7575ef38cb"} Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.483733 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:35:52 crc kubenswrapper[4861]: E0315 00:35:52.484451 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="registry-server" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.484469 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="registry-server" Mar 15 00:35:52 crc kubenswrapper[4861]: E0315 00:35:52.484494 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="extract-utilities" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.484501 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="extract-utilities" Mar 15 00:35:52 crc kubenswrapper[4861]: E0315 00:35:52.484512 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="extract-content" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.484521 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="extract-content" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.484653 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="da1028fa-8876-4981-a6e0-f2f2e5e7d18c" containerName="registry-server" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.485543 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.493112 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.583352 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.583444 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.583863 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4gff\" (UniqueName: \"kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.687318 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.687395 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4gff\" (UniqueName: \"kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.687454 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.687972 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.688028 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.711794 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4gff\" (UniqueName: \"kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff\") pod \"community-operators-8vw66\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:52 crc kubenswrapper[4861]: I0315 00:35:52.810982 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.374321 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:35:54 crc kubenswrapper[4861]: W0315 00:35:54.397021 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce739be9_12f3_4abd_a255_f2554ada5263.slice/crio-62c4617acc3b1ef99d457d7174fbdebe8b658c126f1926d0f6204982b274267d WatchSource:0}: Error finding container 62c4617acc3b1ef99d457d7174fbdebe8b658c126f1926d0f6204982b274267d: Status 404 returned error can't find the container with id 62c4617acc3b1ef99d457d7174fbdebe8b658c126f1926d0f6204982b274267d Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.964889 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48"] Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.966207 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.973881 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.973932 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.973980 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.974020 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-smq8p" Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.980068 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48"] Mar 15 00:35:54 crc kubenswrapper[4861]: I0315 00:35:54.993511 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerStarted","Data":"b9d6ab2fd6c03241e1ab868e73d1c21130e3a70cc547e6b631ea34647811b2c3"} Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.006180 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce739be9-12f3-4abd-a255-f2554ada5263" containerID="fa981084d7675fa0dd9f9208a03b8a7bc5c19c9110a1c8b520a92442f6838be6" exitCode=0 Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.006247 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerDied","Data":"fa981084d7675fa0dd9f9208a03b8a7bc5c19c9110a1c8b520a92442f6838be6"} Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.006305 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerStarted","Data":"62c4617acc3b1ef99d457d7174fbdebe8b658c126f1926d0f6204982b274267d"} Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.131272 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.132249 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8g79\" (UniqueName: \"kubernetes.io/projected/48a91887-8b1e-4e63-b34b-9d169fbde235-kube-api-access-v8g79\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.132302 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/48a91887-8b1e-4e63-b34b-9d169fbde235-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.132360 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/48a91887-8b1e-4e63-b34b-9d169fbde235-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.132393 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.233906 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8g79\" (UniqueName: \"kubernetes.io/projected/48a91887-8b1e-4e63-b34b-9d169fbde235-kube-api-access-v8g79\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.233988 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/48a91887-8b1e-4e63-b34b-9d169fbde235-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.234050 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/48a91887-8b1e-4e63-b34b-9d169fbde235-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.234081 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.234111 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: E0315 00:35:55.234276 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 15 00:35:55 crc kubenswrapper[4861]: E0315 00:35:55.234344 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls podName:48a91887-8b1e-4e63-b34b-9d169fbde235 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:55.734320626 +0000 UTC m=+1749.046518159 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" (UID: "48a91887-8b1e-4e63-b34b-9d169fbde235") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.235015 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/48a91887-8b1e-4e63-b34b-9d169fbde235-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.235731 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/48a91887-8b1e-4e63-b34b-9d169fbde235-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.244332 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.266648 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8g79\" (UniqueName: \"kubernetes.io/projected/48a91887-8b1e-4e63-b34b-9d169fbde235-kube-api-access-v8g79\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: I0315 00:35:55.742643 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:55 crc kubenswrapper[4861]: E0315 00:35:55.742798 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Mar 15 00:35:55 crc kubenswrapper[4861]: E0315 00:35:55.742849 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls podName:48a91887-8b1e-4e63-b34b-9d169fbde235 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:56.742834206 +0000 UTC m=+1750.055031739 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" (UID: "48a91887-8b1e-4e63-b34b-9d169fbde235") : secret "default-cloud1-coll-meter-proxy-tls" not found Mar 15 00:35:56 crc kubenswrapper[4861]: I0315 00:35:56.409427 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:35:56 crc kubenswrapper[4861]: E0315 00:35:56.409909 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:35:56 crc kubenswrapper[4861]: I0315 00:35:56.761062 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:56 crc kubenswrapper[4861]: I0315 00:35:56.844187 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/48a91887-8b1e-4e63-b34b-9d169fbde235-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48\" (UID: \"48a91887-8b1e-4e63-b34b-9d169fbde235\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.023763 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce739be9-12f3-4abd-a255-f2554ada5263" containerID="1078e61b39df6e1c515b30299a5e12416708fb780e575f31d67ae77158319746" exitCode=0 Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.023851 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerDied","Data":"1078e61b39df6e1c515b30299a5e12416708fb780e575f31d67ae77158319746"} Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.091452 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.347548 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48"] Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.738919 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2"] Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.741330 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.744102 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.745456 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.756026 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2"] Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.885949 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.886014 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67cr\" (UniqueName: \"kubernetes.io/projected/97629229-d4e2-48c8-973e-3121ccca2803-kube-api-access-c67cr\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.886059 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97629229-d4e2-48c8-973e-3121ccca2803-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.886098 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97629229-d4e2-48c8-973e-3121ccca2803-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.886128 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.987839 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: E0315 00:35:57.987905 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 15 00:35:57 crc kubenswrapper[4861]: E0315 00:35:57.987991 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls podName:97629229-d4e2-48c8-973e-3121ccca2803 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:58.487971053 +0000 UTC m=+1751.800168586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" (UID: "97629229-d4e2-48c8-973e-3121ccca2803") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.987911 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67cr\" (UniqueName: \"kubernetes.io/projected/97629229-d4e2-48c8-973e-3121ccca2803-kube-api-access-c67cr\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.988250 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97629229-d4e2-48c8-973e-3121ccca2803-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.988424 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97629229-d4e2-48c8-973e-3121ccca2803-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.988504 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.989041 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97629229-d4e2-48c8-973e-3121ccca2803-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.989979 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97629229-d4e2-48c8-973e-3121ccca2803-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:57 crc kubenswrapper[4861]: I0315 00:35:57.997089 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.012499 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67cr\" (UniqueName: \"kubernetes.io/projected/97629229-d4e2-48c8-973e-3121ccca2803-kube-api-access-c67cr\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.036000 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"19034d07fc9fe54fb86b91e9f1b50c2f97ac1934f6c617e0b7270640a5a0264f"} Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.048404 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerStarted","Data":"47dea1582da3377740cedfce65f4b59c2e1afb80538fc7b3cb532f21293bee7a"} Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.055318 4861 generic.go:334] "Generic (PLEG): container finished" podID="931f3bda-5390-448a-b844-f198f756dcfa" containerID="ba0d4c0656927866a201deb6de1d8e1836236cd55b6d541b0f0d2b7575ef38cb" exitCode=0 Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.055487 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerDied","Data":"ba0d4c0656927866a201deb6de1d8e1836236cd55b6d541b0f0d2b7575ef38cb"} Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.059751 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerStarted","Data":"eb23cdf2184dc412ba7eeb3ee9df76d1880b5a134a530842f70482497b86ba88"} Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.070476 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8vw66" podStartSLOduration=3.680603982 podStartE2EDuration="6.070456273s" podCreationTimestamp="2026-03-15 00:35:52 +0000 UTC" firstStartedPulling="2026-03-15 00:35:55.009147857 +0000 UTC m=+1748.321345390" lastFinishedPulling="2026-03-15 00:35:57.399000138 +0000 UTC m=+1750.711197681" observedRunningTime="2026-03-15 00:35:58.069380245 +0000 UTC m=+1751.381577788" watchObservedRunningTime="2026-03-15 00:35:58.070456273 +0000 UTC m=+1751.382653806" Mar 15 00:35:58 crc kubenswrapper[4861]: I0315 00:35:58.503379 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:58 crc kubenswrapper[4861]: E0315 00:35:58.503621 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 15 00:35:58 crc kubenswrapper[4861]: E0315 00:35:58.503721 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls podName:97629229-d4e2-48c8-973e-3121ccca2803 nodeName:}" failed. No retries permitted until 2026-03-15 00:35:59.503696298 +0000 UTC m=+1752.815893831 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" (UID: "97629229-d4e2-48c8-973e-3121ccca2803") : secret "default-cloud1-ceil-meter-proxy-tls" not found Mar 15 00:35:59 crc kubenswrapper[4861]: I0315 00:35:59.523598 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:59 crc kubenswrapper[4861]: I0315 00:35:59.530424 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97629229-d4e2-48c8-973e-3121ccca2803-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2\" (UID: \"97629229-d4e2-48c8-973e-3121ccca2803\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:35:59 crc kubenswrapper[4861]: I0315 00:35:59.558809 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.143584 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558916-sbtft"] Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.144902 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.148478 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.148816 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.149241 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.171325 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558916-sbtft"] Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.236811 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrzp5\" (UniqueName: \"kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5\") pod \"auto-csr-approver-29558916-sbtft\" (UID: \"1b24217a-8fe7-4e28-85d4-8053361942af\") " pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.338204 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrzp5\" (UniqueName: \"kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5\") pod \"auto-csr-approver-29558916-sbtft\" (UID: \"1b24217a-8fe7-4e28-85d4-8053361942af\") " pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.371493 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrzp5\" (UniqueName: \"kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5\") pod \"auto-csr-approver-29558916-sbtft\" (UID: \"1b24217a-8fe7-4e28-85d4-8053361942af\") " pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:00 crc kubenswrapper[4861]: I0315 00:36:00.466034 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.393104 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk"] Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.399187 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.401964 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.402349 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.419308 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk"] Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.554803 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.555038 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.555262 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24fqc\" (UniqueName: \"kubernetes.io/projected/97014776-0e85-4ae2-adde-e6d39175f79c-kube-api-access-24fqc\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.555308 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97014776-0e85-4ae2-adde-e6d39175f79c-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.555395 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97014776-0e85-4ae2-adde-e6d39175f79c-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.656889 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24fqc\" (UniqueName: \"kubernetes.io/projected/97014776-0e85-4ae2-adde-e6d39175f79c-kube-api-access-24fqc\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.656936 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97014776-0e85-4ae2-adde-e6d39175f79c-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.656971 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97014776-0e85-4ae2-adde-e6d39175f79c-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.657296 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.657355 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/97014776-0e85-4ae2-adde-e6d39175f79c-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.657393 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: E0315 00:36:01.657507 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 15 00:36:01 crc kubenswrapper[4861]: E0315 00:36:01.657602 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls podName:97014776-0e85-4ae2-adde-e6d39175f79c nodeName:}" failed. No retries permitted until 2026-03-15 00:36:02.157549946 +0000 UTC m=+1755.469747469 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" (UID: "97014776-0e85-4ae2-adde-e6d39175f79c") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.658487 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/97014776-0e85-4ae2-adde-e6d39175f79c-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.665678 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:01 crc kubenswrapper[4861]: I0315 00:36:01.673217 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24fqc\" (UniqueName: \"kubernetes.io/projected/97014776-0e85-4ae2-adde-e6d39175f79c-kube-api-access-24fqc\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:02 crc kubenswrapper[4861]: I0315 00:36:02.165850 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:02 crc kubenswrapper[4861]: E0315 00:36:02.166160 4861 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Mar 15 00:36:02 crc kubenswrapper[4861]: E0315 00:36:02.166264 4861 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls podName:97014776-0e85-4ae2-adde-e6d39175f79c nodeName:}" failed. No retries permitted until 2026-03-15 00:36:03.166239351 +0000 UTC m=+1756.478436904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" (UID: "97014776-0e85-4ae2-adde-e6d39175f79c") : secret "default-cloud1-sens-meter-proxy-tls" not found Mar 15 00:36:02 crc kubenswrapper[4861]: I0315 00:36:02.811776 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:02 crc kubenswrapper[4861]: I0315 00:36:02.811846 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:02 crc kubenswrapper[4861]: I0315 00:36:02.857842 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:03 crc kubenswrapper[4861]: I0315 00:36:03.164496 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:03 crc kubenswrapper[4861]: I0315 00:36:03.182100 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:03 crc kubenswrapper[4861]: I0315 00:36:03.201425 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/97014776-0e85-4ae2-adde-e6d39175f79c-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk\" (UID: \"97014776-0e85-4ae2-adde-e6d39175f79c\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:03 crc kubenswrapper[4861]: I0315 00:36:03.218601 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" Mar 15 00:36:03 crc kubenswrapper[4861]: I0315 00:36:03.228793 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:36:04 crc kubenswrapper[4861]: I0315 00:36:04.154301 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2"] Mar 15 00:36:05 crc kubenswrapper[4861]: I0315 00:36:05.111883 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8vw66" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="registry-server" containerID="cri-o://47dea1582da3377740cedfce65f4b59c2e1afb80538fc7b3cb532f21293bee7a" gracePeriod=2 Mar 15 00:36:06 crc kubenswrapper[4861]: I0315 00:36:06.120601 4861 generic.go:334] "Generic (PLEG): container finished" podID="ce739be9-12f3-4abd-a255-f2554ada5263" containerID="47dea1582da3377740cedfce65f4b59c2e1afb80538fc7b3cb532f21293bee7a" exitCode=0 Mar 15 00:36:06 crc kubenswrapper[4861]: I0315 00:36:06.120699 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerDied","Data":"47dea1582da3377740cedfce65f4b59c2e1afb80538fc7b3cb532f21293bee7a"} Mar 15 00:36:06 crc kubenswrapper[4861]: I0315 00:36:06.122883 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"c8d5d6e76dea7f29979efcfac1ce876fa748df5aef65852cefd070ec3034b222"} Mar 15 00:36:06 crc kubenswrapper[4861]: I0315 00:36:06.919827 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558916-sbtft"] Mar 15 00:36:06 crc kubenswrapper[4861]: I0315 00:36:06.951031 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.011833 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk"] Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.053477 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities\") pod \"ce739be9-12f3-4abd-a255-f2554ada5263\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.053641 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content\") pod \"ce739be9-12f3-4abd-a255-f2554ada5263\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.053719 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4gff\" (UniqueName: \"kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff\") pod \"ce739be9-12f3-4abd-a255-f2554ada5263\" (UID: \"ce739be9-12f3-4abd-a255-f2554ada5263\") " Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.055033 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities" (OuterVolumeSpecName: "utilities") pod "ce739be9-12f3-4abd-a255-f2554ada5263" (UID: "ce739be9-12f3-4abd-a255-f2554ada5263"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.061277 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff" (OuterVolumeSpecName: "kube-api-access-d4gff") pod "ce739be9-12f3-4abd-a255-f2554ada5263" (UID: "ce739be9-12f3-4abd-a255-f2554ada5263"). InnerVolumeSpecName "kube-api-access-d4gff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.139966 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"289619546aa556af9d483151ba611a2f5d7928cf3c82cb0450a87045a1699b05"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.147639 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558916-sbtft" event={"ID":"1b24217a-8fe7-4e28-85d4-8053361942af","Type":"ContainerStarted","Data":"4ae8cd26de73c55bb34dd607b293da00a1d9278f5b6f38b9bdf9322b052de7a4"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.150004 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"6338a980a9806338123db6b6d9241a1595e9f03a15d73ff275828c8075815807"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.152408 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vw66" event={"ID":"ce739be9-12f3-4abd-a255-f2554ada5263","Type":"ContainerDied","Data":"62c4617acc3b1ef99d457d7174fbdebe8b658c126f1926d0f6204982b274267d"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.152448 4861 scope.go:117] "RemoveContainer" containerID="47dea1582da3377740cedfce65f4b59c2e1afb80538fc7b3cb532f21293bee7a" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.152743 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vw66" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.157710 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4gff\" (UniqueName: \"kubernetes.io/projected/ce739be9-12f3-4abd-a255-f2554ada5263-kube-api-access-d4gff\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.157770 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.172703 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerStarted","Data":"41f6fa8846e74c422d00e880d713af829e2bae6c0c09ae8c917e83afc15e5b0b"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.174200 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"2b2673f780d622c41da81728ef38e8ccb9cf947afcf945519f1bdbd887e628fa"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.181182 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"ef5a360a-3596-48be-8d28-ee75cfee0900","Type":"ContainerStarted","Data":"7394df90e6f52258600cf7d0dbb2f35d2db72aa17533081e857060536b0e940b"} Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.183870 4861 scope.go:117] "RemoveContainer" containerID="1078e61b39df6e1c515b30299a5e12416708fb780e575f31d67ae77158319746" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.218423 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=3.889289724 podStartE2EDuration="42.218401076s" podCreationTimestamp="2026-03-15 00:35:25 +0000 UTC" firstStartedPulling="2026-03-15 00:35:28.417403392 +0000 UTC m=+1721.729600915" lastFinishedPulling="2026-03-15 00:36:06.746514734 +0000 UTC m=+1760.058712267" observedRunningTime="2026-03-15 00:36:07.209807211 +0000 UTC m=+1760.522004754" watchObservedRunningTime="2026-03-15 00:36:07.218401076 +0000 UTC m=+1760.530598609" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.220411 4861 scope.go:117] "RemoveContainer" containerID="fa981084d7675fa0dd9f9208a03b8a7bc5c19c9110a1c8b520a92442f6838be6" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.235073 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce739be9-12f3-4abd-a255-f2554ada5263" (UID: "ce739be9-12f3-4abd-a255-f2554ada5263"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.261279 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce739be9-12f3-4abd-a255-f2554ada5263-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.495398 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:36:07 crc kubenswrapper[4861]: I0315 00:36:07.506317 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8vw66"] Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.050351 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.203842 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"0f004f9e498e08fbc83f70391fc30cf6c8f5c0bc7e9c9c3d0c1a34b7bb19dbeb"} Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.229902 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"7b524c7482d304dfa6c52fb0754a0d1edc3501e9e911a28713ab4f845cdd52c2"} Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.395003 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7"] Mar 15 00:36:08 crc kubenswrapper[4861]: E0315 00:36:08.395426 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="registry-server" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.395442 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="registry-server" Mar 15 00:36:08 crc kubenswrapper[4861]: E0315 00:36:08.395462 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="extract-content" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.395469 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="extract-content" Mar 15 00:36:08 crc kubenswrapper[4861]: E0315 00:36:08.395488 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="extract-utilities" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.395495 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="extract-utilities" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.395669 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" containerName="registry-server" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.396522 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.401909 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.404580 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.406807 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7"] Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.409280 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:36:08 crc kubenswrapper[4861]: E0315 00:36:08.409481 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.587838 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/00278a87-0d72-42d1-aa47-00e60d647563-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.588157 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/00278a87-0d72-42d1-aa47-00e60d647563-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.588180 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8bp\" (UniqueName: \"kubernetes.io/projected/00278a87-0d72-42d1-aa47-00e60d647563-kube-api-access-wf8bp\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.588205 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/00278a87-0d72-42d1-aa47-00e60d647563-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.689752 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/00278a87-0d72-42d1-aa47-00e60d647563-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.689804 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/00278a87-0d72-42d1-aa47-00e60d647563-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.689829 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8bp\" (UniqueName: \"kubernetes.io/projected/00278a87-0d72-42d1-aa47-00e60d647563-kube-api-access-wf8bp\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.689858 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/00278a87-0d72-42d1-aa47-00e60d647563-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.690418 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/00278a87-0d72-42d1-aa47-00e60d647563-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.690711 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/00278a87-0d72-42d1-aa47-00e60d647563-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.705085 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/00278a87-0d72-42d1-aa47-00e60d647563-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.706581 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8bp\" (UniqueName: \"kubernetes.io/projected/00278a87-0d72-42d1-aa47-00e60d647563-kube-api-access-wf8bp\") pod \"default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7\" (UID: \"00278a87-0d72-42d1-aa47-00e60d647563\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:08 crc kubenswrapper[4861]: I0315 00:36:08.727853 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.246523 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7"] Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.255380 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"6090718c06ba3f797460cf2936d21a8c7514a0c246d473a61d5f703dc5273104"} Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.269093 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerStarted","Data":"361067a9a64797a392a11eab3a3a6c10e347ab94cea13b9118bb935b003babaf"} Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.273951 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"cd5cc4c4ffaa5bdbd685ce11f1cf3aae0f12cdf40ddac12c7d981ec353842d95"} Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.285098 4861 generic.go:334] "Generic (PLEG): container finished" podID="1b24217a-8fe7-4e28-85d4-8053361942af" containerID="278a53a72a4bfee42403bf71d89af2f66b89367c0163c28a98249cc164d64705" exitCode=0 Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.285180 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558916-sbtft" event={"ID":"1b24217a-8fe7-4e28-85d4-8053361942af","Type":"ContainerDied","Data":"278a53a72a4bfee42403bf71d89af2f66b89367c0163c28a98249cc164d64705"} Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.420545 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce739be9-12f3-4abd-a255-f2554ada5263" path="/var/lib/kubelet/pods/ce739be9-12f3-4abd-a255-f2554ada5263/volumes" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.720517 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98"] Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.721976 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.724631 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.733101 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/f1aa284f-d724-448f-82b8-d853a24dad74-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.733203 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f1aa284f-d724-448f-82b8-d853a24dad74-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.733260 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f1aa284f-d724-448f-82b8-d853a24dad74-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.733421 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxsrg\" (UniqueName: \"kubernetes.io/projected/f1aa284f-d724-448f-82b8-d853a24dad74-kube-api-access-qxsrg\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.736251 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98"] Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.835337 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f1aa284f-d724-448f-82b8-d853a24dad74-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.835508 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxsrg\" (UniqueName: \"kubernetes.io/projected/f1aa284f-d724-448f-82b8-d853a24dad74-kube-api-access-qxsrg\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.835578 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/f1aa284f-d724-448f-82b8-d853a24dad74-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.835622 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f1aa284f-d724-448f-82b8-d853a24dad74-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.835857 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/f1aa284f-d724-448f-82b8-d853a24dad74-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.836738 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/f1aa284f-d724-448f-82b8-d853a24dad74-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.842767 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/f1aa284f-d724-448f-82b8-d853a24dad74-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:09 crc kubenswrapper[4861]: I0315 00:36:09.852026 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxsrg\" (UniqueName: \"kubernetes.io/projected/f1aa284f-d724-448f-82b8-d853a24dad74-kube-api-access-qxsrg\") pod \"default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98\" (UID: \"f1aa284f-d724-448f-82b8-d853a24dad74\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.048728 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.297583 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"931f3bda-5390-448a-b844-f198f756dcfa","Type":"ContainerStarted","Data":"722b006ad8c586293f9be47e8f41a1691997d22a31bcca2af87f02807d616b40"} Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.304501 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerStarted","Data":"5eab6fbf77fa24029940147b6f944d990a7642b8eb107bcbb89b90786b5cf844"} Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.304875 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerStarted","Data":"417300d45f2d3ca5cd9fab046094af8b6c328c69bb882563e48825ac5323b559"} Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.543124 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=20.445336531 podStartE2EDuration="31.543104535s" podCreationTimestamp="2026-03-15 00:35:39 +0000 UTC" firstStartedPulling="2026-03-15 00:35:58.057333489 +0000 UTC m=+1751.369531022" lastFinishedPulling="2026-03-15 00:36:09.155101493 +0000 UTC m=+1762.467299026" observedRunningTime="2026-03-15 00:36:10.331836537 +0000 UTC m=+1763.644034070" watchObservedRunningTime="2026-03-15 00:36:10.543104535 +0000 UTC m=+1763.855302068" Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.549050 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98"] Mar 15 00:36:10 crc kubenswrapper[4861]: W0315 00:36:10.568458 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1aa284f_d724_448f_82b8_d853a24dad74.slice/crio-1aed7bcb0cf71d61c93f325a0dce2b802a8397e9c332187ab63c4b6bc9eed885 WatchSource:0}: Error finding container 1aed7bcb0cf71d61c93f325a0dce2b802a8397e9c332187ab63c4b6bc9eed885: Status 404 returned error can't find the container with id 1aed7bcb0cf71d61c93f325a0dce2b802a8397e9c332187ab63c4b6bc9eed885 Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.671925 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.863855 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrzp5\" (UniqueName: \"kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5\") pod \"1b24217a-8fe7-4e28-85d4-8053361942af\" (UID: \"1b24217a-8fe7-4e28-85d4-8053361942af\") " Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.871796 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5" (OuterVolumeSpecName: "kube-api-access-qrzp5") pod "1b24217a-8fe7-4e28-85d4-8053361942af" (UID: "1b24217a-8fe7-4e28-85d4-8053361942af"). InnerVolumeSpecName "kube-api-access-qrzp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:36:10 crc kubenswrapper[4861]: I0315 00:36:10.966095 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrzp5\" (UniqueName: \"kubernetes.io/projected/1b24217a-8fe7-4e28-85d4-8053361942af-kube-api-access-qrzp5\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.332875 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerStarted","Data":"1aed7bcb0cf71d61c93f325a0dce2b802a8397e9c332187ab63c4b6bc9eed885"} Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.339113 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558916-sbtft" Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.339491 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558916-sbtft" event={"ID":"1b24217a-8fe7-4e28-85d4-8053361942af","Type":"ContainerDied","Data":"4ae8cd26de73c55bb34dd607b293da00a1d9278f5b6f38b9bdf9322b052de7a4"} Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.341534 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ae8cd26de73c55bb34dd607b293da00a1d9278f5b6f38b9bdf9322b052de7a4" Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.737503 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558910-gjt28"] Mar 15 00:36:11 crc kubenswrapper[4861]: I0315 00:36:11.743246 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558910-gjt28"] Mar 15 00:36:13 crc kubenswrapper[4861]: I0315 00:36:13.050110 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Mar 15 00:36:13 crc kubenswrapper[4861]: I0315 00:36:13.090233 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Mar 15 00:36:13 crc kubenswrapper[4861]: I0315 00:36:13.424226 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89f5a73f-6841-4b40-be0c-31c0e4760f86" path="/var/lib/kubelet/pods/89f5a73f-6841-4b40-be0c-31c0e4760f86/volumes" Mar 15 00:36:13 crc kubenswrapper[4861]: I0315 00:36:13.426451 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Mar 15 00:36:19 crc kubenswrapper[4861]: I0315 00:36:19.408657 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:36:19 crc kubenswrapper[4861]: E0315 00:36:19.409713 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.437757 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"c5fced72354b8a57ac67a487791ec191e8b4bef7a72ca25679b3be3c8b54fece"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.441582 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"6caeb027af0443e1c181bc02442de6ab3f092a317a8a5210a5e02a5029301d75"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.445676 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"18d62524fad07f32c3f1a13009f14933732ddcc1f221c96c88f755d83a4ac783"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.448458 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerStarted","Data":"f5dc660388106e3864b4da79f648b5e896c678777cdc42226bdcac7912c2c656"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.448511 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerStarted","Data":"92d9e1376eb59780a27d3970d84b1fc7400d8d7aaede5ce76e63875470de4bb9"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.453099 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerStarted","Data":"befdbd1309a881827203f7970898be19c7850117e82db70970000bffd353bf75"} Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.488896 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" podStartSLOduration=4.555477248 podStartE2EDuration="26.488859239s" podCreationTimestamp="2026-03-15 00:35:54 +0000 UTC" firstStartedPulling="2026-03-15 00:35:57.359928391 +0000 UTC m=+1750.672125924" lastFinishedPulling="2026-03-15 00:36:19.293310382 +0000 UTC m=+1772.605507915" observedRunningTime="2026-03-15 00:36:20.469372358 +0000 UTC m=+1773.781569921" watchObservedRunningTime="2026-03-15 00:36:20.488859239 +0000 UTC m=+1773.801056812" Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.512029 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" podStartSLOduration=2.412776292 podStartE2EDuration="12.512003369s" podCreationTimestamp="2026-03-15 00:36:08 +0000 UTC" firstStartedPulling="2026-03-15 00:36:09.258679146 +0000 UTC m=+1762.570876679" lastFinishedPulling="2026-03-15 00:36:19.357906223 +0000 UTC m=+1772.670103756" observedRunningTime="2026-03-15 00:36:20.5083536 +0000 UTC m=+1773.820551173" watchObservedRunningTime="2026-03-15 00:36:20.512003369 +0000 UTC m=+1773.824200932" Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.544614 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" podStartSLOduration=2.4140159629999998 podStartE2EDuration="11.544545917s" podCreationTimestamp="2026-03-15 00:36:09 +0000 UTC" firstStartedPulling="2026-03-15 00:36:10.582758746 +0000 UTC m=+1763.894956279" lastFinishedPulling="2026-03-15 00:36:19.7132887 +0000 UTC m=+1773.025486233" observedRunningTime="2026-03-15 00:36:20.54392933 +0000 UTC m=+1773.856126873" watchObservedRunningTime="2026-03-15 00:36:20.544545917 +0000 UTC m=+1773.856743480" Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.580300 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" podStartSLOduration=7.25024856 podStartE2EDuration="19.58026156s" podCreationTimestamp="2026-03-15 00:36:01 +0000 UTC" firstStartedPulling="2026-03-15 00:36:07.028363245 +0000 UTC m=+1760.340560778" lastFinishedPulling="2026-03-15 00:36:19.358376245 +0000 UTC m=+1772.670573778" observedRunningTime="2026-03-15 00:36:20.563827172 +0000 UTC m=+1773.876024715" watchObservedRunningTime="2026-03-15 00:36:20.58026156 +0000 UTC m=+1773.892459133" Mar 15 00:36:20 crc kubenswrapper[4861]: I0315 00:36:20.595704 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" podStartSLOduration=9.514668953 podStartE2EDuration="23.59568022s" podCreationTimestamp="2026-03-15 00:35:57 +0000 UTC" firstStartedPulling="2026-03-15 00:36:05.216304945 +0000 UTC m=+1758.528502488" lastFinishedPulling="2026-03-15 00:36:19.297316182 +0000 UTC m=+1772.609513755" observedRunningTime="2026-03-15 00:36:20.585211965 +0000 UTC m=+1773.897409508" watchObservedRunningTime="2026-03-15 00:36:20.59568022 +0000 UTC m=+1773.907877763" Mar 15 00:36:24 crc kubenswrapper[4861]: I0315 00:36:24.930095 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:36:24 crc kubenswrapper[4861]: I0315 00:36:24.931362 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-7487t" podUID="2ae02148-c9b7-4035-abc2-1f86adce42ec" containerName="default-interconnect" containerID="cri-o://660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6" gracePeriod=30 Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.375023 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.489508 4861 generic.go:334] "Generic (PLEG): container finished" podID="2ae02148-c9b7-4035-abc2-1f86adce42ec" containerID="660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6" exitCode=0 Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.489633 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-7487t" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.489626 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-7487t" event={"ID":"2ae02148-c9b7-4035-abc2-1f86adce42ec","Type":"ContainerDied","Data":"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6"} Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.489715 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-7487t" event={"ID":"2ae02148-c9b7-4035-abc2-1f86adce42ec","Type":"ContainerDied","Data":"b1cf3be3f80a2d911866f0296f5d6b5d6ed77bcfba2b9539a88049314b31575e"} Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.489749 4861 scope.go:117] "RemoveContainer" containerID="660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.493783 4861 generic.go:334] "Generic (PLEG): container finished" podID="48a91887-8b1e-4e63-b34b-9d169fbde235" containerID="0f004f9e498e08fbc83f70391fc30cf6c8f5c0bc7e9c9c3d0c1a34b7bb19dbeb" exitCode=0 Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.493837 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerDied","Data":"0f004f9e498e08fbc83f70391fc30cf6c8f5c0bc7e9c9c3d0c1a34b7bb19dbeb"} Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.494810 4861 scope.go:117] "RemoveContainer" containerID="0f004f9e498e08fbc83f70391fc30cf6c8f5c0bc7e9c9c3d0c1a34b7bb19dbeb" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.495916 4861 generic.go:334] "Generic (PLEG): container finished" podID="97629229-d4e2-48c8-973e-3121ccca2803" containerID="6090718c06ba3f797460cf2936d21a8c7514a0c246d473a61d5f703dc5273104" exitCode=0 Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.495956 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerDied","Data":"6090718c06ba3f797460cf2936d21a8c7514a0c246d473a61d5f703dc5273104"} Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.496533 4861 scope.go:117] "RemoveContainer" containerID="6090718c06ba3f797460cf2936d21a8c7514a0c246d473a61d5f703dc5273104" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523627 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdmp5\" (UniqueName: \"kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523751 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523795 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523828 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523865 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523892 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.523931 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials\") pod \"2ae02148-c9b7-4035-abc2-1f86adce42ec\" (UID: \"2ae02148-c9b7-4035-abc2-1f86adce42ec\") " Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.528661 4861 scope.go:117] "RemoveContainer" containerID="660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.532186 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.538876 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.539227 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: E0315 00:36:25.539250 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6\": container with ID starting with 660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6 not found: ID does not exist" containerID="660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.539311 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6"} err="failed to get container status \"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6\": rpc error: code = NotFound desc = could not find container \"660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6\": container with ID starting with 660fe771534c4da3e9525dd6d3900720cb62711a1b5e690ad0d57dfe6d8eb5a6 not found: ID does not exist" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.550941 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5" (OuterVolumeSpecName: "kube-api-access-wdmp5") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "kube-api-access-wdmp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.551475 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.557069 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.558340 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "2ae02148-c9b7-4035-abc2-1f86adce42ec" (UID: "2ae02148-c9b7-4035-abc2-1f86adce42ec"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629790 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdmp5\" (UniqueName: \"kubernetes.io/projected/2ae02148-c9b7-4035-abc2-1f86adce42ec-kube-api-access-wdmp5\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629838 4861 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629850 4861 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-users\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629863 4861 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/2ae02148-c9b7-4035-abc2-1f86adce42ec-sasl-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629872 4861 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629884 4861 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.629894 4861 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/2ae02148-c9b7-4035-abc2-1f86adce42ec-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.835589 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:36:25 crc kubenswrapper[4861]: I0315 00:36:25.841458 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-7487t"] Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.325249 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-78847"] Mar 15 00:36:26 crc kubenswrapper[4861]: E0315 00:36:26.326119 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b24217a-8fe7-4e28-85d4-8053361942af" containerName="oc" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.326144 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b24217a-8fe7-4e28-85d4-8053361942af" containerName="oc" Mar 15 00:36:26 crc kubenswrapper[4861]: E0315 00:36:26.326171 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae02148-c9b7-4035-abc2-1f86adce42ec" containerName="default-interconnect" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.326180 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae02148-c9b7-4035-abc2-1f86adce42ec" containerName="default-interconnect" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.326303 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b24217a-8fe7-4e28-85d4-8053361942af" containerName="oc" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.326315 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae02148-c9b7-4035-abc2-1f86adce42ec" containerName="default-interconnect" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.326911 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.329462 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.329539 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.333815 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.333869 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.334171 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.340199 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-7fdxt" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.344506 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.347160 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-78847"] Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441622 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441690 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-users\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441727 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgml\" (UniqueName: \"kubernetes.io/projected/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-kube-api-access-xvgml\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441765 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441834 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441895 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-config\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.441945 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.504686 4861 generic.go:334] "Generic (PLEG): container finished" podID="f1aa284f-d724-448f-82b8-d853a24dad74" containerID="92d9e1376eb59780a27d3970d84b1fc7400d8d7aaede5ce76e63875470de4bb9" exitCode=0 Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.504748 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerDied","Data":"92d9e1376eb59780a27d3970d84b1fc7400d8d7aaede5ce76e63875470de4bb9"} Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.505271 4861 scope.go:117] "RemoveContainer" containerID="92d9e1376eb59780a27d3970d84b1fc7400d8d7aaede5ce76e63875470de4bb9" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.509141 4861 generic.go:334] "Generic (PLEG): container finished" podID="00278a87-0d72-42d1-aa47-00e60d647563" containerID="5eab6fbf77fa24029940147b6f944d990a7642b8eb107bcbb89b90786b5cf844" exitCode=0 Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.509196 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerDied","Data":"5eab6fbf77fa24029940147b6f944d990a7642b8eb107bcbb89b90786b5cf844"} Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.509509 4861 scope.go:117] "RemoveContainer" containerID="5eab6fbf77fa24029940147b6f944d990a7642b8eb107bcbb89b90786b5cf844" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.512177 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"f8ca93127ada5d4dda03a6446f036d3374d0264fc52b2af80cd46d1fe6a98191"} Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.514394 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"8e3e62b08305267f356e285a20577b3cbb1e44177762e10b0370becba5a986d0"} Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.517590 4861 generic.go:334] "Generic (PLEG): container finished" podID="97014776-0e85-4ae2-adde-e6d39175f79c" containerID="cd5cc4c4ffaa5bdbd685ce11f1cf3aae0f12cdf40ddac12c7d981ec353842d95" exitCode=0 Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.517623 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerDied","Data":"cd5cc4c4ffaa5bdbd685ce11f1cf3aae0f12cdf40ddac12c7d981ec353842d95"} Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.517966 4861 scope.go:117] "RemoveContainer" containerID="cd5cc4c4ffaa5bdbd685ce11f1cf3aae0f12cdf40ddac12c7d981ec353842d95" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543065 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-config\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543189 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543228 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543252 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-users\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543279 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgml\" (UniqueName: \"kubernetes.io/projected/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-kube-api-access-xvgml\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543322 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.543412 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.545404 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-config\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.555239 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.556321 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.556832 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.557288 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-sasl-users\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.566002 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.577311 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgml\" (UniqueName: \"kubernetes.io/projected/cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2-kube-api-access-xvgml\") pod \"default-interconnect-68864d46cb-78847\" (UID: \"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2\") " pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.657445 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-78847" Mar 15 00:36:26 crc kubenswrapper[4861]: I0315 00:36:26.949912 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-78847"] Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.419481 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae02148-c9b7-4035-abc2-1f86adce42ec" path="/var/lib/kubelet/pods/2ae02148-c9b7-4035-abc2-1f86adce42ec/volumes" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.530002 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-78847" event={"ID":"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2","Type":"ContainerStarted","Data":"25564b5102b73610566cb6eb25b1c01b730a2855b37374b23f1ba15dd35b70ac"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.530108 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-78847" event={"ID":"cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2","Type":"ContainerStarted","Data":"01a816d3618d8e34a7b19292986677cad9a73e1305f0457b1ff8fd90a09671e9"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.533048 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerStarted","Data":"f7c7a3bc14d13b1cddb8990673768d9075762dd545a9a2a89e5279fd15e72239"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.536304 4861 generic.go:334] "Generic (PLEG): container finished" podID="48a91887-8b1e-4e63-b34b-9d169fbde235" containerID="f8ca93127ada5d4dda03a6446f036d3374d0264fc52b2af80cd46d1fe6a98191" exitCode=0 Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.536375 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerDied","Data":"f8ca93127ada5d4dda03a6446f036d3374d0264fc52b2af80cd46d1fe6a98191"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.536437 4861 scope.go:117] "RemoveContainer" containerID="0f004f9e498e08fbc83f70391fc30cf6c8f5c0bc7e9c9c3d0c1a34b7bb19dbeb" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.536930 4861 scope.go:117] "RemoveContainer" containerID="f8ca93127ada5d4dda03a6446f036d3374d0264fc52b2af80cd46d1fe6a98191" Mar 15 00:36:27 crc kubenswrapper[4861]: E0315 00:36:27.537486 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48_service-telemetry(48a91887-8b1e-4e63-b34b-9d169fbde235)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" podUID="48a91887-8b1e-4e63-b34b-9d169fbde235" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.539726 4861 generic.go:334] "Generic (PLEG): container finished" podID="97629229-d4e2-48c8-973e-3121ccca2803" containerID="8e3e62b08305267f356e285a20577b3cbb1e44177762e10b0370becba5a986d0" exitCode=0 Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.539789 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerDied","Data":"8e3e62b08305267f356e285a20577b3cbb1e44177762e10b0370becba5a986d0"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.540169 4861 scope.go:117] "RemoveContainer" containerID="8e3e62b08305267f356e285a20577b3cbb1e44177762e10b0370becba5a986d0" Mar 15 00:36:27 crc kubenswrapper[4861]: E0315 00:36:27.540391 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2_service-telemetry(97629229-d4e2-48c8-973e-3121ccca2803)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" podUID="97629229-d4e2-48c8-973e-3121ccca2803" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.554335 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-78847" podStartSLOduration=3.5543216490000002 podStartE2EDuration="3.554321649s" podCreationTimestamp="2026-03-15 00:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-15 00:36:27.550963268 +0000 UTC m=+1780.863160811" watchObservedRunningTime="2026-03-15 00:36:27.554321649 +0000 UTC m=+1780.866519192" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.602858 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"169554340ff8a4ab2444bc0c8e7dbf4556b4485c79d8227c0b5461f7671c9e43"} Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.611572 4861 scope.go:117] "RemoveContainer" containerID="6090718c06ba3f797460cf2936d21a8c7514a0c246d473a61d5f703dc5273104" Mar 15 00:36:27 crc kubenswrapper[4861]: I0315 00:36:27.611786 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerStarted","Data":"7997adde680608f01580dedb32960e5dce3eadff13e1ef9f31c15fd6eded3f1e"} Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.622724 4861 generic.go:334] "Generic (PLEG): container finished" podID="97014776-0e85-4ae2-adde-e6d39175f79c" containerID="169554340ff8a4ab2444bc0c8e7dbf4556b4485c79d8227c0b5461f7671c9e43" exitCode=0 Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.622819 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerDied","Data":"169554340ff8a4ab2444bc0c8e7dbf4556b4485c79d8227c0b5461f7671c9e43"} Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.622936 4861 scope.go:117] "RemoveContainer" containerID="cd5cc4c4ffaa5bdbd685ce11f1cf3aae0f12cdf40ddac12c7d981ec353842d95" Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.623699 4861 scope.go:117] "RemoveContainer" containerID="169554340ff8a4ab2444bc0c8e7dbf4556b4485c79d8227c0b5461f7671c9e43" Mar 15 00:36:28 crc kubenswrapper[4861]: E0315 00:36:28.623984 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk_service-telemetry(97014776-0e85-4ae2-adde-e6d39175f79c)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" podUID="97014776-0e85-4ae2-adde-e6d39175f79c" Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.628048 4861 generic.go:334] "Generic (PLEG): container finished" podID="f1aa284f-d724-448f-82b8-d853a24dad74" containerID="7997adde680608f01580dedb32960e5dce3eadff13e1ef9f31c15fd6eded3f1e" exitCode=0 Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.628174 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerDied","Data":"7997adde680608f01580dedb32960e5dce3eadff13e1ef9f31c15fd6eded3f1e"} Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.628921 4861 scope.go:117] "RemoveContainer" containerID="7997adde680608f01580dedb32960e5dce3eadff13e1ef9f31c15fd6eded3f1e" Mar 15 00:36:28 crc kubenswrapper[4861]: E0315 00:36:28.629318 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98_service-telemetry(f1aa284f-d724-448f-82b8-d853a24dad74)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" podUID="f1aa284f-d724-448f-82b8-d853a24dad74" Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.630672 4861 generic.go:334] "Generic (PLEG): container finished" podID="00278a87-0d72-42d1-aa47-00e60d647563" containerID="f7c7a3bc14d13b1cddb8990673768d9075762dd545a9a2a89e5279fd15e72239" exitCode=0 Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.630727 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerDied","Data":"f7c7a3bc14d13b1cddb8990673768d9075762dd545a9a2a89e5279fd15e72239"} Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.631458 4861 scope.go:117] "RemoveContainer" containerID="f7c7a3bc14d13b1cddb8990673768d9075762dd545a9a2a89e5279fd15e72239" Mar 15 00:36:28 crc kubenswrapper[4861]: E0315 00:36:28.631738 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7_service-telemetry(00278a87-0d72-42d1-aa47-00e60d647563)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" podUID="00278a87-0d72-42d1-aa47-00e60d647563" Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.761480 4861 scope.go:117] "RemoveContainer" containerID="92d9e1376eb59780a27d3970d84b1fc7400d8d7aaede5ce76e63875470de4bb9" Mar 15 00:36:28 crc kubenswrapper[4861]: I0315 00:36:28.801410 4861 scope.go:117] "RemoveContainer" containerID="5eab6fbf77fa24029940147b6f944d990a7642b8eb107bcbb89b90786b5cf844" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.205654 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.207322 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.210488 4861 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.221540 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.253543 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.326706 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcdvw\" (UniqueName: \"kubernetes.io/projected/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-kube-api-access-pcdvw\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.326935 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-qdr-test-config\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.327052 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.428883 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcdvw\" (UniqueName: \"kubernetes.io/projected/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-kube-api-access-pcdvw\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.428952 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-qdr-test-config\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.429014 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.430019 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-qdr-test-config\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.438432 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.450170 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcdvw\" (UniqueName: \"kubernetes.io/projected/47ba8d94-9c61-42db-af30-f3f1a3ea2a79-kube-api-access-pcdvw\") pod \"qdr-test\" (UID: \"47ba8d94-9c61-42db-af30-f3f1a3ea2a79\") " pod="service-telemetry/qdr-test" Mar 15 00:36:30 crc kubenswrapper[4861]: I0315 00:36:30.568961 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Mar 15 00:36:31 crc kubenswrapper[4861]: I0315 00:36:31.042717 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Mar 15 00:36:31 crc kubenswrapper[4861]: W0315 00:36:31.054278 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47ba8d94_9c61_42db_af30_f3f1a3ea2a79.slice/crio-0cef33e544b8abf498f24ea266de0d74b40bad8c33f7c5544e4607ad2abaf2b5 WatchSource:0}: Error finding container 0cef33e544b8abf498f24ea266de0d74b40bad8c33f7c5544e4607ad2abaf2b5: Status 404 returned error can't find the container with id 0cef33e544b8abf498f24ea266de0d74b40bad8c33f7c5544e4607ad2abaf2b5 Mar 15 00:36:31 crc kubenswrapper[4861]: I0315 00:36:31.702815 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"47ba8d94-9c61-42db-af30-f3f1a3ea2a79","Type":"ContainerStarted","Data":"0cef33e544b8abf498f24ea266de0d74b40bad8c33f7c5544e4607ad2abaf2b5"} Mar 15 00:36:34 crc kubenswrapper[4861]: I0315 00:36:34.409316 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:36:34 crc kubenswrapper[4861]: E0315 00:36:34.409548 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:36:39 crc kubenswrapper[4861]: I0315 00:36:39.768887 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"47ba8d94-9c61-42db-af30-f3f1a3ea2a79","Type":"ContainerStarted","Data":"b97fa346b72b256b15ebc178c84df984f9a3c0ed02ef0591a59e59518f03c6c6"} Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.071949 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=1.7701189670000002 podStartE2EDuration="10.071921915s" podCreationTimestamp="2026-03-15 00:36:30 +0000 UTC" firstStartedPulling="2026-03-15 00:36:31.05699993 +0000 UTC m=+1784.369197463" lastFinishedPulling="2026-03-15 00:36:39.358802858 +0000 UTC m=+1792.671000411" observedRunningTime="2026-03-15 00:36:39.791830821 +0000 UTC m=+1793.104028354" watchObservedRunningTime="2026-03-15 00:36:40.071921915 +0000 UTC m=+1793.384119448" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.073945 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-xrrf6"] Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.100042 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.103536 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.104492 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.104726 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.111906 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.112285 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.113734 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.123198 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-xrrf6"] Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.196519 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.196600 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4wl\" (UniqueName: \"kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.196637 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.196671 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.197162 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.197313 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.197491 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299389 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299467 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299518 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299539 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4wl\" (UniqueName: \"kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299588 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299654 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.299716 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.300419 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.301246 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.301317 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.301348 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.302020 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.302940 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.336120 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4wl\" (UniqueName: \"kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl\") pod \"stf-smoketest-smoke1-xrrf6\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.409054 4861 scope.go:117] "RemoveContainer" containerID="f8ca93127ada5d4dda03a6446f036d3374d0264fc52b2af80cd46d1fe6a98191" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.426463 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.454185 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.455815 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.476678 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.607399 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx87n\" (UniqueName: \"kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n\") pod \"curl\" (UID: \"e49045e1-7e53-4146-94e8-9c3b61c3769b\") " pod="service-telemetry/curl" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.688052 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-xrrf6"] Mar 15 00:36:40 crc kubenswrapper[4861]: W0315 00:36:40.695368 4861 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97027e85_bc7b_44fc_b622_84b2cb5817bd.slice/crio-1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8 WatchSource:0}: Error finding container 1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8: Status 404 returned error can't find the container with id 1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8 Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.709699 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx87n\" (UniqueName: \"kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n\") pod \"curl\" (UID: \"e49045e1-7e53-4146-94e8-9c3b61c3769b\") " pod="service-telemetry/curl" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.736289 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx87n\" (UniqueName: \"kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n\") pod \"curl\" (UID: \"e49045e1-7e53-4146-94e8-9c3b61c3769b\") " pod="service-telemetry/curl" Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.779595 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerStarted","Data":"1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8"} Mar 15 00:36:40 crc kubenswrapper[4861]: I0315 00:36:40.839429 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.104318 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.409697 4861 scope.go:117] "RemoveContainer" containerID="169554340ff8a4ab2444bc0c8e7dbf4556b4485c79d8227c0b5461f7671c9e43" Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.409965 4861 scope.go:117] "RemoveContainer" containerID="f7c7a3bc14d13b1cddb8990673768d9075762dd545a9a2a89e5279fd15e72239" Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.410076 4861 scope.go:117] "RemoveContainer" containerID="7997adde680608f01580dedb32960e5dce3eadff13e1ef9f31c15fd6eded3f1e" Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.413611 4861 scope.go:117] "RemoveContainer" containerID="8e3e62b08305267f356e285a20577b3cbb1e44177762e10b0370becba5a986d0" Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.798518 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48" event={"ID":"48a91887-8b1e-4e63-b34b-9d169fbde235","Type":"ContainerStarted","Data":"4eced70963b371edbd27c87bd1438165f9f3393866dd83448cb578cb186cfcc8"} Mar 15 00:36:41 crc kubenswrapper[4861]: I0315 00:36:41.804917 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e49045e1-7e53-4146-94e8-9c3b61c3769b","Type":"ContainerStarted","Data":"584bde1543407c07a7d2c7f2da99efe392aed7ab0120eb9c1f08d4a203c576d4"} Mar 15 00:36:42 crc kubenswrapper[4861]: I0315 00:36:42.818071 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2" event={"ID":"97629229-d4e2-48c8-973e-3121ccca2803","Type":"ContainerStarted","Data":"6b258163c9935af50939e0d36637c7b098080a85b0df1e88bb478bcb8a8f8764"} Mar 15 00:36:42 crc kubenswrapper[4861]: I0315 00:36:42.822493 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk" event={"ID":"97014776-0e85-4ae2-adde-e6d39175f79c","Type":"ContainerStarted","Data":"bbc3c0c8d43af6e0e93112462c6cf4965970176f77e721ee70587eed7e0452c2"} Mar 15 00:36:42 crc kubenswrapper[4861]: I0315 00:36:42.825449 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98" event={"ID":"f1aa284f-d724-448f-82b8-d853a24dad74","Type":"ContainerStarted","Data":"891e3d49f6464ed3adff14ca0155c9a3dbbcb92709c8643d7579d533b44266d2"} Mar 15 00:36:42 crc kubenswrapper[4861]: I0315 00:36:42.833874 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7" event={"ID":"00278a87-0d72-42d1-aa47-00e60d647563","Type":"ContainerStarted","Data":"8bc80fbe529da633db656eba1771abe3c7c5d91afa563f81ebce106758f897fb"} Mar 15 00:36:45 crc kubenswrapper[4861]: I0315 00:36:45.409633 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:36:45 crc kubenswrapper[4861]: E0315 00:36:45.410120 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:36:45 crc kubenswrapper[4861]: I0315 00:36:45.865239 4861 generic.go:334] "Generic (PLEG): container finished" podID="e49045e1-7e53-4146-94e8-9c3b61c3769b" containerID="f8b2ee0cb7963018693ca3d79a8eb1982d2422e6e6bbe9589877b2b1e7f4d7f4" exitCode=0 Mar 15 00:36:45 crc kubenswrapper[4861]: I0315 00:36:45.865297 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e49045e1-7e53-4146-94e8-9c3b61c3769b","Type":"ContainerDied","Data":"f8b2ee0cb7963018693ca3d79a8eb1982d2422e6e6bbe9589877b2b1e7f4d7f4"} Mar 15 00:36:48 crc kubenswrapper[4861]: I0315 00:36:48.914330 4861 scope.go:117] "RemoveContainer" containerID="c80cf9f3d8c68c3730133799c00a540ce5a01f1d6d6e4a710cde1ba836f2e2ce" Mar 15 00:36:48 crc kubenswrapper[4861]: I0315 00:36:48.920066 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"e49045e1-7e53-4146-94e8-9c3b61c3769b","Type":"ContainerDied","Data":"584bde1543407c07a7d2c7f2da99efe392aed7ab0120eb9c1f08d4a203c576d4"} Mar 15 00:36:48 crc kubenswrapper[4861]: I0315 00:36:48.920128 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="584bde1543407c07a7d2c7f2da99efe392aed7ab0120eb9c1f08d4a203c576d4" Mar 15 00:36:48 crc kubenswrapper[4861]: I0315 00:36:48.932413 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.056423 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx87n\" (UniqueName: \"kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n\") pod \"e49045e1-7e53-4146-94e8-9c3b61c3769b\" (UID: \"e49045e1-7e53-4146-94e8-9c3b61c3769b\") " Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.073699 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n" (OuterVolumeSpecName: "kube-api-access-zx87n") pod "e49045e1-7e53-4146-94e8-9c3b61c3769b" (UID: "e49045e1-7e53-4146-94e8-9c3b61c3769b"). InnerVolumeSpecName "kube-api-access-zx87n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.089605 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_e49045e1-7e53-4146-94e8-9c3b61c3769b/curl/0.log" Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.158181 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx87n\" (UniqueName: \"kubernetes.io/projected/e49045e1-7e53-4146-94e8-9c3b61c3769b-kube-api-access-zx87n\") on node \"crc\" DevicePath \"\"" Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.291858 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-qs59c_32144527-d295-4a45-ba18-ba6f391d60a9/prometheus-webhook-snmp/0.log" Mar 15 00:36:49 crc kubenswrapper[4861]: I0315 00:36:49.928836 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Mar 15 00:36:54 crc kubenswrapper[4861]: I0315 00:36:54.981523 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerStarted","Data":"4062b9e8c986763915a2ff400b5443847e5990fa3a21b73ef0350f7bd48dc2fa"} Mar 15 00:37:00 crc kubenswrapper[4861]: I0315 00:37:00.409268 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:37:00 crc kubenswrapper[4861]: E0315 00:37:00.412422 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:37:01 crc kubenswrapper[4861]: I0315 00:37:01.038540 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerStarted","Data":"cbf9093bee13588e13ef62bef5cd5246a5cc8809337405944b2d57b0eec139a8"} Mar 15 00:37:01 crc kubenswrapper[4861]: I0315 00:37:01.070105 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" podStartSLOduration=1.927051059 podStartE2EDuration="21.070076452s" podCreationTimestamp="2026-03-15 00:36:40 +0000 UTC" firstStartedPulling="2026-03-15 00:36:40.69802577 +0000 UTC m=+1794.010223303" lastFinishedPulling="2026-03-15 00:36:59.841051153 +0000 UTC m=+1813.153248696" observedRunningTime="2026-03-15 00:37:01.06600313 +0000 UTC m=+1814.378200703" watchObservedRunningTime="2026-03-15 00:37:01.070076452 +0000 UTC m=+1814.382274015" Mar 15 00:37:13 crc kubenswrapper[4861]: I0315 00:37:13.410268 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:37:13 crc kubenswrapper[4861]: E0315 00:37:13.410999 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:37:19 crc kubenswrapper[4861]: I0315 00:37:19.471707 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-qs59c_32144527-d295-4a45-ba18-ba6f391d60a9/prometheus-webhook-snmp/0.log" Mar 15 00:37:27 crc kubenswrapper[4861]: I0315 00:37:27.419787 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:37:27 crc kubenswrapper[4861]: E0315 00:37:27.421159 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:37:28 crc kubenswrapper[4861]: I0315 00:37:28.338603 4861 generic.go:334] "Generic (PLEG): container finished" podID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerID="4062b9e8c986763915a2ff400b5443847e5990fa3a21b73ef0350f7bd48dc2fa" exitCode=0 Mar 15 00:37:28 crc kubenswrapper[4861]: I0315 00:37:28.338667 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerDied","Data":"4062b9e8c986763915a2ff400b5443847e5990fa3a21b73ef0350f7bd48dc2fa"} Mar 15 00:37:28 crc kubenswrapper[4861]: I0315 00:37:28.339424 4861 scope.go:117] "RemoveContainer" containerID="4062b9e8c986763915a2ff400b5443847e5990fa3a21b73ef0350f7bd48dc2fa" Mar 15 00:37:32 crc kubenswrapper[4861]: I0315 00:37:32.382936 4861 generic.go:334] "Generic (PLEG): container finished" podID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerID="cbf9093bee13588e13ef62bef5cd5246a5cc8809337405944b2d57b0eec139a8" exitCode=0 Mar 15 00:37:32 crc kubenswrapper[4861]: I0315 00:37:32.383051 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerDied","Data":"cbf9093bee13588e13ef62bef5cd5246a5cc8809337405944b2d57b0eec139a8"} Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.767165 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911606 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911654 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911697 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911766 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911798 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4wl\" (UniqueName: \"kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911838 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.911870 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher\") pod \"97027e85-bc7b-44fc-b622-84b2cb5817bd\" (UID: \"97027e85-bc7b-44fc-b622-84b2cb5817bd\") " Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.920055 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl" (OuterVolumeSpecName: "kube-api-access-qd4wl") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "kube-api-access-qd4wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.935480 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.936660 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.943958 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.944444 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.945635 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:33 crc kubenswrapper[4861]: I0315 00:37:33.948150 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "97027e85-bc7b-44fc-b622-84b2cb5817bd" (UID: "97027e85-bc7b-44fc-b622-84b2cb5817bd"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014329 4861 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014469 4861 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-sensubility-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014543 4861 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014627 4861 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-healthcheck-log\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014653 4861 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-collectd-config\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014678 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4wl\" (UniqueName: \"kubernetes.io/projected/97027e85-bc7b-44fc-b622-84b2cb5817bd-kube-api-access-qd4wl\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.014753 4861 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/97027e85-bc7b-44fc-b622-84b2cb5817bd-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.403072 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" event={"ID":"97027e85-bc7b-44fc-b622-84b2cb5817bd","Type":"ContainerDied","Data":"1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8"} Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.403133 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf92595028ffcd85bb2122745735e770ef66b77c5adae1440419a27d9b367a8" Mar 15 00:37:34 crc kubenswrapper[4861]: I0315 00:37:34.403196 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-xrrf6" Mar 15 00:37:35 crc kubenswrapper[4861]: I0315 00:37:35.697212 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-xrrf6_97027e85-bc7b-44fc-b622-84b2cb5817bd/smoketest-collectd/0.log" Mar 15 00:37:35 crc kubenswrapper[4861]: I0315 00:37:35.957760 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-xrrf6_97027e85-bc7b-44fc-b622-84b2cb5817bd/smoketest-ceilometer/0.log" Mar 15 00:37:36 crc kubenswrapper[4861]: I0315 00:37:36.219895 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-78847_cd15c11c-65a2-4784-8fb3-0a1eccf2e3c2/default-interconnect/0.log" Mar 15 00:37:36 crc kubenswrapper[4861]: I0315 00:37:36.455320 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48_48a91887-8b1e-4e63-b34b-9d169fbde235/bridge/2.log" Mar 15 00:37:36 crc kubenswrapper[4861]: I0315 00:37:36.689131 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-ngw48_48a91887-8b1e-4e63-b34b-9d169fbde235/sg-core/0.log" Mar 15 00:37:36 crc kubenswrapper[4861]: I0315 00:37:36.933472 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7_00278a87-0d72-42d1-aa47-00e60d647563/bridge/2.log" Mar 15 00:37:37 crc kubenswrapper[4861]: I0315 00:37:37.195119 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6fdbbbb78-8cbj7_00278a87-0d72-42d1-aa47-00e60d647563/sg-core/0.log" Mar 15 00:37:37 crc kubenswrapper[4861]: I0315 00:37:37.485269 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2_97629229-d4e2-48c8-973e-3121ccca2803/bridge/2.log" Mar 15 00:37:37 crc kubenswrapper[4861]: I0315 00:37:37.734106 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-ctgk2_97629229-d4e2-48c8-973e-3121ccca2803/sg-core/0.log" Mar 15 00:37:37 crc kubenswrapper[4861]: I0315 00:37:37.961402 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98_f1aa284f-d724-448f-82b8-d853a24dad74/bridge/2.log" Mar 15 00:37:38 crc kubenswrapper[4861]: I0315 00:37:38.198899 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-69977f5f84-s9m98_f1aa284f-d724-448f-82b8-d853a24dad74/sg-core/0.log" Mar 15 00:37:38 crc kubenswrapper[4861]: I0315 00:37:38.590511 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk_97014776-0e85-4ae2-adde-e6d39175f79c/bridge/2.log" Mar 15 00:37:38 crc kubenswrapper[4861]: I0315 00:37:38.802299 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-dwqrk_97014776-0e85-4ae2-adde-e6d39175f79c/sg-core/0.log" Mar 15 00:37:40 crc kubenswrapper[4861]: I0315 00:37:40.408731 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:37:40 crc kubenswrapper[4861]: E0315 00:37:40.409309 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:37:42 crc kubenswrapper[4861]: I0315 00:37:42.413830 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-5fcfd8dffb-6sd8k_1ea04940-819c-401c-842f-86b7b48ee8eb/operator/0.log" Mar 15 00:37:42 crc kubenswrapper[4861]: I0315 00:37:42.650349 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_ef5a360a-3596-48be-8d28-ee75cfee0900/prometheus/0.log" Mar 15 00:37:42 crc kubenswrapper[4861]: I0315 00:37:42.964921 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_47c92463-1387-4e36-a0a4-00ced68121fd/elasticsearch/0.log" Mar 15 00:37:43 crc kubenswrapper[4861]: I0315 00:37:43.254796 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-qs59c_32144527-d295-4a45-ba18-ba6f391d60a9/prometheus-webhook-snmp/0.log" Mar 15 00:37:43 crc kubenswrapper[4861]: I0315 00:37:43.482969 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_931f3bda-5390-448a-b844-f198f756dcfa/alertmanager/0.log" Mar 15 00:37:51 crc kubenswrapper[4861]: I0315 00:37:51.409641 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:37:51 crc kubenswrapper[4861]: E0315 00:37:51.410534 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:37:54 crc kubenswrapper[4861]: I0315 00:37:54.136158 4861 scope.go:117] "RemoveContainer" containerID="45b662408f36b7f9e4a466d8b7d1ac0a72d03af330284ea8b47ad01cbad69089" Mar 15 00:37:54 crc kubenswrapper[4861]: I0315 00:37:54.199111 4861 scope.go:117] "RemoveContainer" containerID="1bf4e329bde3aef2737a7208e9fa79e473602819bd693a8fca561f207be6d179" Mar 15 00:37:58 crc kubenswrapper[4861]: I0315 00:37:58.591065 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-86bfdfcc6-jwwzs_ee2860e1-04df-4782-9bf0-0baf205e595d/operator/0.log" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145348 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558918-jlckk"] Mar 15 00:38:00 crc kubenswrapper[4861]: E0315 00:38:00.145708 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e49045e1-7e53-4146-94e8-9c3b61c3769b" containerName="curl" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145723 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e49045e1-7e53-4146-94e8-9c3b61c3769b" containerName="curl" Mar 15 00:38:00 crc kubenswrapper[4861]: E0315 00:38:00.145742 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-ceilometer" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145749 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-ceilometer" Mar 15 00:38:00 crc kubenswrapper[4861]: E0315 00:38:00.145758 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-collectd" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145766 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-collectd" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145887 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-collectd" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145900 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="e49045e1-7e53-4146-94e8-9c3b61c3769b" containerName="curl" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.145908 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="97027e85-bc7b-44fc-b622-84b2cb5817bd" containerName="smoketest-ceilometer" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.146374 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.149968 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.149968 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.151923 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.155191 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558918-jlckk"] Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.258965 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh85d\" (UniqueName: \"kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d\") pod \"auto-csr-approver-29558918-jlckk\" (UID: \"e0e2742f-8dbf-4a35-abf8-34044e539708\") " pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.361224 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh85d\" (UniqueName: \"kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d\") pod \"auto-csr-approver-29558918-jlckk\" (UID: \"e0e2742f-8dbf-4a35-abf8-34044e539708\") " pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.385013 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh85d\" (UniqueName: \"kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d\") pod \"auto-csr-approver-29558918-jlckk\" (UID: \"e0e2742f-8dbf-4a35-abf8-34044e539708\") " pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.469902 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:00 crc kubenswrapper[4861]: I0315 00:38:00.955152 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558918-jlckk"] Mar 15 00:38:01 crc kubenswrapper[4861]: I0315 00:38:01.666923 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558918-jlckk" event={"ID":"e0e2742f-8dbf-4a35-abf8-34044e539708","Type":"ContainerStarted","Data":"f25043e35e4f649d05453ba54deb93962f0d80934a3add72f52b96e811977eaf"} Mar 15 00:38:02 crc kubenswrapper[4861]: I0315 00:38:02.248313 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-5fcfd8dffb-6sd8k_1ea04940-819c-401c-842f-86b7b48ee8eb/operator/0.log" Mar 15 00:38:02 crc kubenswrapper[4861]: I0315 00:38:02.527071 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_47ba8d94-9c61-42db-af30-f3f1a3ea2a79/qdr/0.log" Mar 15 00:38:02 crc kubenswrapper[4861]: I0315 00:38:02.680698 4861 generic.go:334] "Generic (PLEG): container finished" podID="e0e2742f-8dbf-4a35-abf8-34044e539708" containerID="d31ca13f7cad7abf3edd3a1adbce4cb60a73c05e09afe41737e2ea7648e40146" exitCode=0 Mar 15 00:38:02 crc kubenswrapper[4861]: I0315 00:38:02.680754 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558918-jlckk" event={"ID":"e0e2742f-8dbf-4a35-abf8-34044e539708","Type":"ContainerDied","Data":"d31ca13f7cad7abf3edd3a1adbce4cb60a73c05e09afe41737e2ea7648e40146"} Mar 15 00:38:03 crc kubenswrapper[4861]: I0315 00:38:03.408820 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:38:03 crc kubenswrapper[4861]: E0315 00:38:03.409469 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.064813 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.120755 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh85d\" (UniqueName: \"kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d\") pod \"e0e2742f-8dbf-4a35-abf8-34044e539708\" (UID: \"e0e2742f-8dbf-4a35-abf8-34044e539708\") " Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.130350 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d" (OuterVolumeSpecName: "kube-api-access-fh85d") pod "e0e2742f-8dbf-4a35-abf8-34044e539708" (UID: "e0e2742f-8dbf-4a35-abf8-34044e539708"). InnerVolumeSpecName "kube-api-access-fh85d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.223455 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh85d\" (UniqueName: \"kubernetes.io/projected/e0e2742f-8dbf-4a35-abf8-34044e539708-kube-api-access-fh85d\") on node \"crc\" DevicePath \"\"" Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.700135 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558918-jlckk" event={"ID":"e0e2742f-8dbf-4a35-abf8-34044e539708","Type":"ContainerDied","Data":"f25043e35e4f649d05453ba54deb93962f0d80934a3add72f52b96e811977eaf"} Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.700196 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f25043e35e4f649d05453ba54deb93962f0d80934a3add72f52b96e811977eaf" Mar 15 00:38:04 crc kubenswrapper[4861]: I0315 00:38:04.700265 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558918-jlckk" Mar 15 00:38:05 crc kubenswrapper[4861]: I0315 00:38:05.159062 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558912-qwsn4"] Mar 15 00:38:05 crc kubenswrapper[4861]: I0315 00:38:05.171003 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558912-qwsn4"] Mar 15 00:38:05 crc kubenswrapper[4861]: I0315 00:38:05.424628 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67bf6356-73f3-4534-84a0-6f5ea23d7c77" path="/var/lib/kubelet/pods/67bf6356-73f3-4534-84a0-6f5ea23d7c77/volumes" Mar 15 00:38:17 crc kubenswrapper[4861]: I0315 00:38:17.418046 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:38:17 crc kubenswrapper[4861]: E0315 00:38:17.420038 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.257637 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-z4klz/must-gather-7xkwd"] Mar 15 00:38:26 crc kubenswrapper[4861]: E0315 00:38:26.260515 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e2742f-8dbf-4a35-abf8-34044e539708" containerName="oc" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.260543 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e2742f-8dbf-4a35-abf8-34044e539708" containerName="oc" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.260815 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e2742f-8dbf-4a35-abf8-34044e539708" containerName="oc" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.261719 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.271360 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-z4klz"/"openshift-service-ca.crt" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.271973 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-z4klz"/"kube-root-ca.crt" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.273027 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-z4klz"/"default-dockercfg-qvtk7" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.299992 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-z4klz/must-gather-7xkwd"] Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.348792 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.348883 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4m9b\" (UniqueName: \"kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.387466 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.397636 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.425412 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.450831 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.451123 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4m9b\" (UniqueName: \"kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.451229 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.472977 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4m9b\" (UniqueName: \"kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b\") pod \"must-gather-7xkwd\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.552711 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psqnm\" (UniqueName: \"kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm\") pod \"infrawatch-operators-g8mjg\" (UID: \"be5eeb37-afc1-4cc6-818d-bd48384246d6\") " pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.582633 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.654729 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psqnm\" (UniqueName: \"kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm\") pod \"infrawatch-operators-g8mjg\" (UID: \"be5eeb37-afc1-4cc6-818d-bd48384246d6\") " pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.679029 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psqnm\" (UniqueName: \"kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm\") pod \"infrawatch-operators-g8mjg\" (UID: \"be5eeb37-afc1-4cc6-818d-bd48384246d6\") " pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.727210 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.818048 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-z4klz/must-gather-7xkwd"] Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.827395 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.936070 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z4klz/must-gather-7xkwd" event={"ID":"3e20d5df-6887-4cc4-ae4f-4fcf5e145976","Type":"ContainerStarted","Data":"c3d7a17b65cd5d1bdf4a62ffbfe6f5c2f6f0baa34f2d08261ceffcca145b621e"} Mar 15 00:38:26 crc kubenswrapper[4861]: I0315 00:38:26.965746 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:27 crc kubenswrapper[4861]: I0315 00:38:27.967781 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8mjg" event={"ID":"be5eeb37-afc1-4cc6-818d-bd48384246d6","Type":"ContainerStarted","Data":"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef"} Mar 15 00:38:27 crc kubenswrapper[4861]: I0315 00:38:27.968143 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8mjg" event={"ID":"be5eeb37-afc1-4cc6-818d-bd48384246d6","Type":"ContainerStarted","Data":"24caeac73b6d309e33704c145cd04ef54885f66b13a3df9c70f204b4ff900ddb"} Mar 15 00:38:28 crc kubenswrapper[4861]: I0315 00:38:28.001363 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-g8mjg" podStartSLOduration=1.8965871349999999 podStartE2EDuration="2.001341771s" podCreationTimestamp="2026-03-15 00:38:26 +0000 UTC" firstStartedPulling="2026-03-15 00:38:26.972728814 +0000 UTC m=+1900.284926347" lastFinishedPulling="2026-03-15 00:38:27.07748345 +0000 UTC m=+1900.389680983" observedRunningTime="2026-03-15 00:38:27.993384504 +0000 UTC m=+1901.305582037" watchObservedRunningTime="2026-03-15 00:38:28.001341771 +0000 UTC m=+1901.313539304" Mar 15 00:38:29 crc kubenswrapper[4861]: I0315 00:38:29.410375 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:38:29 crc kubenswrapper[4861]: E0315 00:38:29.410797 4861 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-r8m45_openshift-machine-config-operator(5442c197-f16b-462d-b7ef-f509c8b1e858)\"" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.049769 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z4klz/must-gather-7xkwd" event={"ID":"3e20d5df-6887-4cc4-ae4f-4fcf5e145976","Type":"ContainerStarted","Data":"920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf"} Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.050897 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z4klz/must-gather-7xkwd" event={"ID":"3e20d5df-6887-4cc4-ae4f-4fcf5e145976","Type":"ContainerStarted","Data":"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307"} Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.081323 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-z4klz/must-gather-7xkwd" podStartSLOduration=2.388303507 podStartE2EDuration="10.081291991s" podCreationTimestamp="2026-03-15 00:38:26 +0000 UTC" firstStartedPulling="2026-03-15 00:38:26.827184867 +0000 UTC m=+1900.139382400" lastFinishedPulling="2026-03-15 00:38:34.520173321 +0000 UTC m=+1907.832370884" observedRunningTime="2026-03-15 00:38:36.076445339 +0000 UTC m=+1909.388642902" watchObservedRunningTime="2026-03-15 00:38:36.081291991 +0000 UTC m=+1909.393489554" Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.727697 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.728136 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:36 crc kubenswrapper[4861]: I0315 00:38:36.789722 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:37 crc kubenswrapper[4861]: I0315 00:38:37.112761 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:37 crc kubenswrapper[4861]: I0315 00:38:37.183161 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:39 crc kubenswrapper[4861]: I0315 00:38:39.077275 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-g8mjg" podUID="be5eeb37-afc1-4cc6-818d-bd48384246d6" containerName="registry-server" containerID="cri-o://012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef" gracePeriod=2 Mar 15 00:38:39 crc kubenswrapper[4861]: I0315 00:38:39.491970 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:39 crc kubenswrapper[4861]: I0315 00:38:39.634021 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psqnm\" (UniqueName: \"kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm\") pod \"be5eeb37-afc1-4cc6-818d-bd48384246d6\" (UID: \"be5eeb37-afc1-4cc6-818d-bd48384246d6\") " Mar 15 00:38:39 crc kubenswrapper[4861]: I0315 00:38:39.641106 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm" (OuterVolumeSpecName: "kube-api-access-psqnm") pod "be5eeb37-afc1-4cc6-818d-bd48384246d6" (UID: "be5eeb37-afc1-4cc6-818d-bd48384246d6"). InnerVolumeSpecName "kube-api-access-psqnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:38:39 crc kubenswrapper[4861]: I0315 00:38:39.735820 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psqnm\" (UniqueName: \"kubernetes.io/projected/be5eeb37-afc1-4cc6-818d-bd48384246d6-kube-api-access-psqnm\") on node \"crc\" DevicePath \"\"" Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.091408 4861 generic.go:334] "Generic (PLEG): container finished" podID="be5eeb37-afc1-4cc6-818d-bd48384246d6" containerID="012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef" exitCode=0 Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.091477 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8mjg" event={"ID":"be5eeb37-afc1-4cc6-818d-bd48384246d6","Type":"ContainerDied","Data":"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef"} Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.091502 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-g8mjg" Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.091535 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-g8mjg" event={"ID":"be5eeb37-afc1-4cc6-818d-bd48384246d6","Type":"ContainerDied","Data":"24caeac73b6d309e33704c145cd04ef54885f66b13a3df9c70f204b4ff900ddb"} Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.091602 4861 scope.go:117] "RemoveContainer" containerID="012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef" Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.145394 4861 scope.go:117] "RemoveContainer" containerID="012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef" Mar 15 00:38:40 crc kubenswrapper[4861]: E0315 00:38:40.147153 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef\": container with ID starting with 012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef not found: ID does not exist" containerID="012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef" Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.147236 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef"} err="failed to get container status \"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef\": rpc error: code = NotFound desc = could not find container \"012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef\": container with ID starting with 012d199a9c718acf0205dd361ec33da44c7e332c777c95d767bd4c11cdf45aef not found: ID does not exist" Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.149525 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:40 crc kubenswrapper[4861]: I0315 00:38:40.163615 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-g8mjg"] Mar 15 00:38:41 crc kubenswrapper[4861]: I0315 00:38:41.425520 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be5eeb37-afc1-4cc6-818d-bd48384246d6" path="/var/lib/kubelet/pods/be5eeb37-afc1-4cc6-818d-bd48384246d6/volumes" Mar 15 00:38:44 crc kubenswrapper[4861]: I0315 00:38:44.409623 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:38:45 crc kubenswrapper[4861]: I0315 00:38:45.147102 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"af3dcb954577b27412b25e29bfa5b62ebeb80345ebae068292497a42159b13e0"} Mar 15 00:38:54 crc kubenswrapper[4861]: I0315 00:38:54.283061 4861 scope.go:117] "RemoveContainer" containerID="ee7849619683fa8b50258e9053abc551cfc871a5c8b83b1374a8f0e39a22dc34" Mar 15 00:38:54 crc kubenswrapper[4861]: I0315 00:38:54.337728 4861 scope.go:117] "RemoveContainer" containerID="3ebd11baefc4351d7d192681acc127e4ed22ca5a6378f5b998144cf8cc6dffb3" Mar 15 00:38:54 crc kubenswrapper[4861]: I0315 00:38:54.408715 4861 scope.go:117] "RemoveContainer" containerID="38f3b796ab2d462f3ad9f920452defd44ed7fded5c839d3e5745e9cfdf6784a8" Mar 15 00:39:28 crc kubenswrapper[4861]: I0315 00:39:28.493129 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-g2qt2_58100d65-ee1f-4e57-9140-5519af53b4c6/control-plane-machine-set-operator/0.log" Mar 15 00:39:28 crc kubenswrapper[4861]: I0315 00:39:28.693384 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fbnkg_6ef26501-9849-4bd4-8976-38cc771e868e/kube-rbac-proxy/0.log" Mar 15 00:39:28 crc kubenswrapper[4861]: I0315 00:39:28.747886 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fbnkg_6ef26501-9849-4bd4-8976-38cc771e868e/machine-api-operator/0.log" Mar 15 00:39:44 crc kubenswrapper[4861]: I0315 00:39:44.780084 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-d8l4v_8197885f-6fb5-4b68-9d32-1c39c74d220f/cert-manager-controller/0.log" Mar 15 00:39:44 crc kubenswrapper[4861]: I0315 00:39:44.969935 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-g5bzq_febd875f-8a0e-43ab-a41c-d555e44bb04d/cert-manager-cainjector/0.log" Mar 15 00:39:45 crc kubenswrapper[4861]: I0315 00:39:45.033607 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-v2qct_85ede584-f9ce-419c-8807-f415ecb45d4c/cert-manager-webhook/0.log" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.457050 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:39:53 crc kubenswrapper[4861]: E0315 00:39:53.458026 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5eeb37-afc1-4cc6-818d-bd48384246d6" containerName="registry-server" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.458041 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5eeb37-afc1-4cc6-818d-bd48384246d6" containerName="registry-server" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.458226 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5eeb37-afc1-4cc6-818d-bd48384246d6" containerName="registry-server" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.465394 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.470084 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.534038 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkqzr\" (UniqueName: \"kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.534184 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.534683 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.636967 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkqzr\" (UniqueName: \"kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.637065 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.637142 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.637774 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.638377 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.667780 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkqzr\" (UniqueName: \"kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr\") pod \"redhat-operators-lcn2g\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:53 crc kubenswrapper[4861]: I0315 00:39:53.790355 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:39:54 crc kubenswrapper[4861]: I0315 00:39:54.113542 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:39:54 crc kubenswrapper[4861]: I0315 00:39:54.768796 4861 generic.go:334] "Generic (PLEG): container finished" podID="26e31bf3-5263-4693-92a4-06de5118324d" containerID="7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b" exitCode=0 Mar 15 00:39:54 crc kubenswrapper[4861]: I0315 00:39:54.768860 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerDied","Data":"7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b"} Mar 15 00:39:54 crc kubenswrapper[4861]: I0315 00:39:54.769149 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerStarted","Data":"13c6970a831da2bacbcf86e29f80aa06f6d39e640bfdd6d2fa12df57755fae55"} Mar 15 00:39:55 crc kubenswrapper[4861]: I0315 00:39:55.778183 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerStarted","Data":"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487"} Mar 15 00:39:56 crc kubenswrapper[4861]: I0315 00:39:56.790579 4861 generic.go:334] "Generic (PLEG): container finished" podID="26e31bf3-5263-4693-92a4-06de5118324d" containerID="46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487" exitCode=0 Mar 15 00:39:56 crc kubenswrapper[4861]: I0315 00:39:56.790704 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerDied","Data":"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487"} Mar 15 00:39:57 crc kubenswrapper[4861]: I0315 00:39:57.802104 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerStarted","Data":"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96"} Mar 15 00:39:57 crc kubenswrapper[4861]: I0315 00:39:57.826180 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lcn2g" podStartSLOduration=2.381889768 podStartE2EDuration="4.82615771s" podCreationTimestamp="2026-03-15 00:39:53 +0000 UTC" firstStartedPulling="2026-03-15 00:39:54.770653248 +0000 UTC m=+1988.082850781" lastFinishedPulling="2026-03-15 00:39:57.21492115 +0000 UTC m=+1990.527118723" observedRunningTime="2026-03-15 00:39:57.821060371 +0000 UTC m=+1991.133257914" watchObservedRunningTime="2026-03-15 00:39:57.82615771 +0000 UTC m=+1991.138355263" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.148031 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558920-6j4rv"] Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.150353 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.153326 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.153698 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.153823 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.158982 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558920-6j4rv"] Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.235992 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj88s\" (UniqueName: \"kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s\") pod \"auto-csr-approver-29558920-6j4rv\" (UID: \"825a559d-acb6-4b33-b671-ab3ae95ebf12\") " pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.338047 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj88s\" (UniqueName: \"kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s\") pod \"auto-csr-approver-29558920-6j4rv\" (UID: \"825a559d-acb6-4b33-b671-ab3ae95ebf12\") " pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.368500 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj88s\" (UniqueName: \"kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s\") pod \"auto-csr-approver-29558920-6j4rv\" (UID: \"825a559d-acb6-4b33-b671-ab3ae95ebf12\") " pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.474740 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.794377 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558920-6j4rv"] Mar 15 00:40:00 crc kubenswrapper[4861]: I0315 00:40:00.824089 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" event={"ID":"825a559d-acb6-4b33-b671-ab3ae95ebf12","Type":"ContainerStarted","Data":"2974ea0134a01f559b18c43f8e6b967a2af967b7066178274a99e3a4e67305bc"} Mar 15 00:40:02 crc kubenswrapper[4861]: I0315 00:40:02.845357 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" event={"ID":"825a559d-acb6-4b33-b671-ab3ae95ebf12","Type":"ContainerStarted","Data":"3579a87ee40fd48eeb61936c31fcda2a907ad47cd4990de2fc5522f0a82d278d"} Mar 15 00:40:02 crc kubenswrapper[4861]: I0315 00:40:02.867408 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" podStartSLOduration=1.7581653130000001 podStartE2EDuration="2.867390166s" podCreationTimestamp="2026-03-15 00:40:00 +0000 UTC" firstStartedPulling="2026-03-15 00:40:00.805853056 +0000 UTC m=+1994.118050589" lastFinishedPulling="2026-03-15 00:40:01.915077909 +0000 UTC m=+1995.227275442" observedRunningTime="2026-03-15 00:40:02.865539916 +0000 UTC m=+1996.177737449" watchObservedRunningTime="2026-03-15 00:40:02.867390166 +0000 UTC m=+1996.179587699" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.525369 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-x6mgw_57e95c2a-a805-46c7-b30e-5e583a122421/prometheus-operator/0.log" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.678134 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7_1dc0d506-b1fd-4326-9417-67d37744ed0b/prometheus-operator-admission-webhook/0.log" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.759169 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7_01397ae2-5789-4479-8670-ffea619a6e15/prometheus-operator-admission-webhook/0.log" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.790647 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.790816 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.854818 4861 generic.go:334] "Generic (PLEG): container finished" podID="825a559d-acb6-4b33-b671-ab3ae95ebf12" containerID="3579a87ee40fd48eeb61936c31fcda2a907ad47cd4990de2fc5522f0a82d278d" exitCode=0 Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.854893 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" event={"ID":"825a559d-acb6-4b33-b671-ab3ae95ebf12","Type":"ContainerDied","Data":"3579a87ee40fd48eeb61936c31fcda2a907ad47cd4990de2fc5522f0a82d278d"} Mar 15 00:40:03 crc kubenswrapper[4861]: I0315 00:40:03.920352 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-j58pz_dee11996-b156-48ac-ac7c-cf3df3bad8e5/operator/0.log" Mar 15 00:40:04 crc kubenswrapper[4861]: I0315 00:40:04.047494 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-qr9rk_41a1f7ae-f37f-4d4f-a12b-927e525730c8/perses-operator/0.log" Mar 15 00:40:04 crc kubenswrapper[4861]: I0315 00:40:04.833209 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lcn2g" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" probeResult="failure" output=< Mar 15 00:40:04 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:40:04 crc kubenswrapper[4861]: > Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.117091 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.215952 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj88s\" (UniqueName: \"kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s\") pod \"825a559d-acb6-4b33-b671-ab3ae95ebf12\" (UID: \"825a559d-acb6-4b33-b671-ab3ae95ebf12\") " Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.222188 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s" (OuterVolumeSpecName: "kube-api-access-rj88s") pod "825a559d-acb6-4b33-b671-ab3ae95ebf12" (UID: "825a559d-acb6-4b33-b671-ab3ae95ebf12"). InnerVolumeSpecName "kube-api-access-rj88s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.318314 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj88s\" (UniqueName: \"kubernetes.io/projected/825a559d-acb6-4b33-b671-ab3ae95ebf12-kube-api-access-rj88s\") on node \"crc\" DevicePath \"\"" Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.872693 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" event={"ID":"825a559d-acb6-4b33-b671-ab3ae95ebf12","Type":"ContainerDied","Data":"2974ea0134a01f559b18c43f8e6b967a2af967b7066178274a99e3a4e67305bc"} Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.873090 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2974ea0134a01f559b18c43f8e6b967a2af967b7066178274a99e3a4e67305bc" Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.872776 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558920-6j4rv" Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.937106 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558914-69rrq"] Mar 15 00:40:05 crc kubenswrapper[4861]: I0315 00:40:05.942171 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558914-69rrq"] Mar 15 00:40:07 crc kubenswrapper[4861]: I0315 00:40:07.422044 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0d01275-16e4-416d-bfba-6a950f5fcb51" path="/var/lib/kubelet/pods/b0d01275-16e4-416d-bfba-6a950f5fcb51/volumes" Mar 15 00:40:14 crc kubenswrapper[4861]: I0315 00:40:14.846333 4861 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lcn2g" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" probeResult="failure" output=< Mar 15 00:40:14 crc kubenswrapper[4861]: timeout: failed to connect service ":50051" within 1s Mar 15 00:40:14 crc kubenswrapper[4861]: > Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.345176 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/util/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.578748 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/pull/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.579384 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/util/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.602874 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/pull/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.767537 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/util/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.770838 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/extract/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.815025 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8f5ss6b_17925b27-9408-4067-970e-de025397edf2/pull/0.log" Mar 15 00:40:20 crc kubenswrapper[4861]: I0315 00:40:20.957262 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/util/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.135806 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/util/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.152999 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/pull/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.160855 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/pull/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.406809 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/pull/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.425291 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/extract/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.453888 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39en7gzs_2c747e57-e783-45d0-a353-521096017d15/util/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.600744 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/util/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.772620 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/util/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.810019 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/pull/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.841625 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/pull/0.log" Mar 15 00:40:21 crc kubenswrapper[4861]: I0315 00:40:21.995064 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/util/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.019177 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/pull/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.066189 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5622wr_fc5af859-0f30-4bf8-8eb3-8101cee89857/extract/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.303344 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/util/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.497478 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/util/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.499379 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/pull/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.525327 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/pull/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.714604 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/pull/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.715636 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/util/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.738299 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08q7whn_4e23db65-ec89-427d-aaa4-21d9c00779fd/extract/0.log" Mar 15 00:40:22 crc kubenswrapper[4861]: I0315 00:40:22.887112 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-utilities/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.126951 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-utilities/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.126980 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.129470 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.268822 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-utilities/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.324173 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.522186 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-utilities/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.662919 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pzwlb_47cc9481-9f7a-47f2-933a-93bb592e1bdf/registry-server/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.679570 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-utilities/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.679667 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.786978 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.835313 4861 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.888646 4861 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.923183 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-content/0.log" Mar 15 00:40:23 crc kubenswrapper[4861]: I0315 00:40:23.933019 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.052536 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-smdks_d7fbc705-3f2f-47d2-9b12-c308b061a5db/marketplace-operator/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.080346 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.188278 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.400767 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-content/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.412311 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-content/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.415971 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qzm9v_f5a24ae0-68ec-4e10-9592-fa33d3380453/registry-server/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.417905 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.543362 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.557199 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/extract-content/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.603186 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lcn2g_26e31bf3-5263-4693-92a4-06de5118324d/registry-server/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.619483 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.789461 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-utilities/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.819977 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-content/0.log" Mar 15 00:40:24 crc kubenswrapper[4861]: I0315 00:40:24.825858 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-content/0.log" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.009542 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-content/0.log" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.012524 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/extract-utilities/0.log" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.037335 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lcn2g" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" containerID="cri-o://fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96" gracePeriod=2 Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.456762 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.459134 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w995g_09972fc0-e7fc-4cd4-87d0-50d858ed9e51/registry-server/0.log" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.549432 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkqzr\" (UniqueName: \"kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr\") pod \"26e31bf3-5263-4693-92a4-06de5118324d\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.549512 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content\") pod \"26e31bf3-5263-4693-92a4-06de5118324d\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.549545 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities\") pod \"26e31bf3-5263-4693-92a4-06de5118324d\" (UID: \"26e31bf3-5263-4693-92a4-06de5118324d\") " Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.551044 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities" (OuterVolumeSpecName: "utilities") pod "26e31bf3-5263-4693-92a4-06de5118324d" (UID: "26e31bf3-5263-4693-92a4-06de5118324d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.556264 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr" (OuterVolumeSpecName: "kube-api-access-bkqzr") pod "26e31bf3-5263-4693-92a4-06de5118324d" (UID: "26e31bf3-5263-4693-92a4-06de5118324d"). InnerVolumeSpecName "kube-api-access-bkqzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.652127 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkqzr\" (UniqueName: \"kubernetes.io/projected/26e31bf3-5263-4693-92a4-06de5118324d-kube-api-access-bkqzr\") on node \"crc\" DevicePath \"\"" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.652205 4861 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-utilities\") on node \"crc\" DevicePath \"\"" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.692014 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "26e31bf3-5263-4693-92a4-06de5118324d" (UID: "26e31bf3-5263-4693-92a4-06de5118324d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:40:25 crc kubenswrapper[4861]: I0315 00:40:25.754150 4861 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/26e31bf3-5263-4693-92a4-06de5118324d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.055884 4861 generic.go:334] "Generic (PLEG): container finished" podID="26e31bf3-5263-4693-92a4-06de5118324d" containerID="fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96" exitCode=0 Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.055963 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lcn2g" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.056005 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerDied","Data":"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96"} Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.056372 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lcn2g" event={"ID":"26e31bf3-5263-4693-92a4-06de5118324d","Type":"ContainerDied","Data":"13c6970a831da2bacbcf86e29f80aa06f6d39e640bfdd6d2fa12df57755fae55"} Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.056405 4861 scope.go:117] "RemoveContainer" containerID="fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.090805 4861 scope.go:117] "RemoveContainer" containerID="46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.091796 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.098890 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lcn2g"] Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.123721 4861 scope.go:117] "RemoveContainer" containerID="7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.150490 4861 scope.go:117] "RemoveContainer" containerID="fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96" Mar 15 00:40:26 crc kubenswrapper[4861]: E0315 00:40:26.151054 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96\": container with ID starting with fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96 not found: ID does not exist" containerID="fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.151124 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96"} err="failed to get container status \"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96\": rpc error: code = NotFound desc = could not find container \"fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96\": container with ID starting with fd4e011413a7398010858f9700eec6f8191a746690cc320f9b1d4df8af121d96 not found: ID does not exist" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.151167 4861 scope.go:117] "RemoveContainer" containerID="46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487" Mar 15 00:40:26 crc kubenswrapper[4861]: E0315 00:40:26.151763 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487\": container with ID starting with 46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487 not found: ID does not exist" containerID="46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.151806 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487"} err="failed to get container status \"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487\": rpc error: code = NotFound desc = could not find container \"46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487\": container with ID starting with 46df754305de8ff68533d409516d9d4c592e454f908f61749aed8b1804583487 not found: ID does not exist" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.151836 4861 scope.go:117] "RemoveContainer" containerID="7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b" Mar 15 00:40:26 crc kubenswrapper[4861]: E0315 00:40:26.152221 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b\": container with ID starting with 7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b not found: ID does not exist" containerID="7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b" Mar 15 00:40:26 crc kubenswrapper[4861]: I0315 00:40:26.152296 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b"} err="failed to get container status \"7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b\": rpc error: code = NotFound desc = could not find container \"7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b\": container with ID starting with 7569909e80afeb1d8f6de8175679e425431ae0e03754933e201cce00fbe27c0b not found: ID does not exist" Mar 15 00:40:27 crc kubenswrapper[4861]: I0315 00:40:27.418800 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e31bf3-5263-4693-92a4-06de5118324d" path="/var/lib/kubelet/pods/26e31bf3-5263-4693-92a4-06de5118324d/volumes" Mar 15 00:40:39 crc kubenswrapper[4861]: I0315 00:40:39.837615 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86d6bc8c95-zr5w7_01397ae2-5789-4479-8670-ffea619a6e15/prometheus-operator-admission-webhook/0.log" Mar 15 00:40:39 crc kubenswrapper[4861]: I0315 00:40:39.838671 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-x6mgw_57e95c2a-a805-46c7-b30e-5e583a122421/prometheus-operator/0.log" Mar 15 00:40:39 crc kubenswrapper[4861]: I0315 00:40:39.877536 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86d6bc8c95-gtlp7_1dc0d506-b1fd-4326-9417-67d37744ed0b/prometheus-operator-admission-webhook/0.log" Mar 15 00:40:40 crc kubenswrapper[4861]: I0315 00:40:40.002814 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-j58pz_dee11996-b156-48ac-ac7c-cf3df3bad8e5/operator/0.log" Mar 15 00:40:40 crc kubenswrapper[4861]: I0315 00:40:40.015010 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-qr9rk_41a1f7ae-f37f-4d4f-a12b-927e525730c8/perses-operator/0.log" Mar 15 00:40:54 crc kubenswrapper[4861]: I0315 00:40:54.526640 4861 scope.go:117] "RemoveContainer" containerID="29a66d0c13c295ed5b878cb51639d59c20028afe3237e7e37425415742411364" Mar 15 00:41:02 crc kubenswrapper[4861]: I0315 00:41:02.282526 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:41:02 crc kubenswrapper[4861]: I0315 00:41:02.283348 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:41:31 crc kubenswrapper[4861]: I0315 00:41:31.755832 4861 generic.go:334] "Generic (PLEG): container finished" podID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerID="c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307" exitCode=0 Mar 15 00:41:31 crc kubenswrapper[4861]: I0315 00:41:31.755998 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-z4klz/must-gather-7xkwd" event={"ID":"3e20d5df-6887-4cc4-ae4f-4fcf5e145976","Type":"ContainerDied","Data":"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307"} Mar 15 00:41:31 crc kubenswrapper[4861]: I0315 00:41:31.757309 4861 scope.go:117] "RemoveContainer" containerID="c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307" Mar 15 00:41:32 crc kubenswrapper[4861]: I0315 00:41:32.002425 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z4klz_must-gather-7xkwd_3e20d5df-6887-4cc4-ae4f-4fcf5e145976/gather/0.log" Mar 15 00:41:32 crc kubenswrapper[4861]: I0315 00:41:32.281504 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:41:32 crc kubenswrapper[4861]: I0315 00:41:32.282013 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.045953 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-z4klz/must-gather-7xkwd"] Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.047067 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-z4klz/must-gather-7xkwd" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="copy" containerID="cri-o://920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf" gracePeriod=2 Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.052408 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-z4klz/must-gather-7xkwd"] Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.513243 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z4klz_must-gather-7xkwd_3e20d5df-6887-4cc4-ae4f-4fcf5e145976/copy/0.log" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.514286 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.591433 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output\") pod \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.591629 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4m9b\" (UniqueName: \"kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b\") pod \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\" (UID: \"3e20d5df-6887-4cc4-ae4f-4fcf5e145976\") " Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.614689 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b" (OuterVolumeSpecName: "kube-api-access-m4m9b") pod "3e20d5df-6887-4cc4-ae4f-4fcf5e145976" (UID: "3e20d5df-6887-4cc4-ae4f-4fcf5e145976"). InnerVolumeSpecName "kube-api-access-m4m9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.679263 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3e20d5df-6887-4cc4-ae4f-4fcf5e145976" (UID: "3e20d5df-6887-4cc4-ae4f-4fcf5e145976"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.693828 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4m9b\" (UniqueName: \"kubernetes.io/projected/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-kube-api-access-m4m9b\") on node \"crc\" DevicePath \"\"" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.693861 4861 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3e20d5df-6887-4cc4-ae4f-4fcf5e145976-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.843391 4861 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-z4klz_must-gather-7xkwd_3e20d5df-6887-4cc4-ae4f-4fcf5e145976/copy/0.log" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.843960 4861 generic.go:334] "Generic (PLEG): container finished" podID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerID="920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf" exitCode=143 Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.844023 4861 scope.go:117] "RemoveContainer" containerID="920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.844183 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-z4klz/must-gather-7xkwd" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.904372 4861 scope.go:117] "RemoveContainer" containerID="c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.977631 4861 scope.go:117] "RemoveContainer" containerID="920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf" Mar 15 00:41:39 crc kubenswrapper[4861]: E0315 00:41:39.978364 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf\": container with ID starting with 920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf not found: ID does not exist" containerID="920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.978457 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf"} err="failed to get container status \"920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf\": rpc error: code = NotFound desc = could not find container \"920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf\": container with ID starting with 920cba41b2c1ea199f678f22b31f1f68fb9e308e125d725318ccc262bd8ab6cf not found: ID does not exist" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.978518 4861 scope.go:117] "RemoveContainer" containerID="c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307" Mar 15 00:41:39 crc kubenswrapper[4861]: E0315 00:41:39.979371 4861 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307\": container with ID starting with c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307 not found: ID does not exist" containerID="c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307" Mar 15 00:41:39 crc kubenswrapper[4861]: I0315 00:41:39.979449 4861 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307"} err="failed to get container status \"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307\": rpc error: code = NotFound desc = could not find container \"c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307\": container with ID starting with c05349d8fc9e97a8fe6ee17ba8841d06e3f977485b69619a4f0dc4284d895307 not found: ID does not exist" Mar 15 00:41:41 crc kubenswrapper[4861]: I0315 00:41:41.425114 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" path="/var/lib/kubelet/pods/3e20d5df-6887-4cc4-ae4f-4fcf5e145976/volumes" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.176258 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558922-5fc8h"] Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177711 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="extract-content" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177736 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="extract-content" Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177766 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="extract-utilities" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177784 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="extract-utilities" Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177801 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177816 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177861 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="copy" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177874 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="copy" Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177903 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="gather" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177915 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="gather" Mar 15 00:42:00 crc kubenswrapper[4861]: E0315 00:42:00.177947 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825a559d-acb6-4b33-b671-ab3ae95ebf12" containerName="oc" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.177967 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="825a559d-acb6-4b33-b671-ab3ae95ebf12" containerName="oc" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.187119 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e31bf3-5263-4693-92a4-06de5118324d" containerName="registry-server" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.187193 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="825a559d-acb6-4b33-b671-ab3ae95ebf12" containerName="oc" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.187228 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="gather" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.187248 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e20d5df-6887-4cc4-ae4f-4fcf5e145976" containerName="copy" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.189603 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.195165 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.195910 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.196201 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.196251 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558922-5fc8h"] Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.342208 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjz5j\" (UniqueName: \"kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j\") pod \"auto-csr-approver-29558922-5fc8h\" (UID: \"71643876-7f20-406a-8b63-d2e665cd166c\") " pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.444804 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjz5j\" (UniqueName: \"kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j\") pod \"auto-csr-approver-29558922-5fc8h\" (UID: \"71643876-7f20-406a-8b63-d2e665cd166c\") " pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.482451 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjz5j\" (UniqueName: \"kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j\") pod \"auto-csr-approver-29558922-5fc8h\" (UID: \"71643876-7f20-406a-8b63-d2e665cd166c\") " pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.530775 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:00 crc kubenswrapper[4861]: I0315 00:42:00.822592 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558922-5fc8h"] Mar 15 00:42:01 crc kubenswrapper[4861]: I0315 00:42:01.065140 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" event={"ID":"71643876-7f20-406a-8b63-d2e665cd166c","Type":"ContainerStarted","Data":"2d6fd0c8cabfbba8997a5c4047044a54ce0704f11e46d2578684673a6ab00cec"} Mar 15 00:42:02 crc kubenswrapper[4861]: I0315 00:42:02.281627 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:42:02 crc kubenswrapper[4861]: I0315 00:42:02.282207 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:42:02 crc kubenswrapper[4861]: I0315 00:42:02.282289 4861 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" Mar 15 00:42:02 crc kubenswrapper[4861]: I0315 00:42:02.283215 4861 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af3dcb954577b27412b25e29bfa5b62ebeb80345ebae068292497a42159b13e0"} pod="openshift-machine-config-operator/machine-config-daemon-r8m45" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 15 00:42:02 crc kubenswrapper[4861]: I0315 00:42:02.283324 4861 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" containerID="cri-o://af3dcb954577b27412b25e29bfa5b62ebeb80345ebae068292497a42159b13e0" gracePeriod=600 Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.096179 4861 generic.go:334] "Generic (PLEG): container finished" podID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerID="af3dcb954577b27412b25e29bfa5b62ebeb80345ebae068292497a42159b13e0" exitCode=0 Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.096300 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerDied","Data":"af3dcb954577b27412b25e29bfa5b62ebeb80345ebae068292497a42159b13e0"} Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.096623 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" event={"ID":"5442c197-f16b-462d-b7ef-f509c8b1e858","Type":"ContainerStarted","Data":"0eee3cedac529ea7d7d33ea3eb5f590cb944b8915987155b78c4a1f7337a2622"} Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.096657 4861 scope.go:117] "RemoveContainer" containerID="e9082542155c58ca1b535cd0854058f0828d628879373798a11394443b3a3187" Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.102842 4861 generic.go:334] "Generic (PLEG): container finished" podID="71643876-7f20-406a-8b63-d2e665cd166c" containerID="1d4932dd75ced0b0815f9557d3ae6c55e036c0a493c2bf35bb834125b35cd376" exitCode=0 Mar 15 00:42:03 crc kubenswrapper[4861]: I0315 00:42:03.102880 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" event={"ID":"71643876-7f20-406a-8b63-d2e665cd166c","Type":"ContainerDied","Data":"1d4932dd75ced0b0815f9557d3ae6c55e036c0a493c2bf35bb834125b35cd376"} Mar 15 00:42:04 crc kubenswrapper[4861]: I0315 00:42:04.450673 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:04 crc kubenswrapper[4861]: I0315 00:42:04.609605 4861 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjz5j\" (UniqueName: \"kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j\") pod \"71643876-7f20-406a-8b63-d2e665cd166c\" (UID: \"71643876-7f20-406a-8b63-d2e665cd166c\") " Mar 15 00:42:04 crc kubenswrapper[4861]: I0315 00:42:04.618645 4861 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j" (OuterVolumeSpecName: "kube-api-access-hjz5j") pod "71643876-7f20-406a-8b63-d2e665cd166c" (UID: "71643876-7f20-406a-8b63-d2e665cd166c"). InnerVolumeSpecName "kube-api-access-hjz5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 15 00:42:04 crc kubenswrapper[4861]: I0315 00:42:04.711953 4861 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjz5j\" (UniqueName: \"kubernetes.io/projected/71643876-7f20-406a-8b63-d2e665cd166c-kube-api-access-hjz5j\") on node \"crc\" DevicePath \"\"" Mar 15 00:42:05 crc kubenswrapper[4861]: I0315 00:42:05.146536 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" event={"ID":"71643876-7f20-406a-8b63-d2e665cd166c","Type":"ContainerDied","Data":"2d6fd0c8cabfbba8997a5c4047044a54ce0704f11e46d2578684673a6ab00cec"} Mar 15 00:42:05 crc kubenswrapper[4861]: I0315 00:42:05.147011 4861 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d6fd0c8cabfbba8997a5c4047044a54ce0704f11e46d2578684673a6ab00cec" Mar 15 00:42:05 crc kubenswrapper[4861]: I0315 00:42:05.146648 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558922-5fc8h" Mar 15 00:42:05 crc kubenswrapper[4861]: I0315 00:42:05.517745 4861 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29558916-sbtft"] Mar 15 00:42:05 crc kubenswrapper[4861]: I0315 00:42:05.528433 4861 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29558916-sbtft"] Mar 15 00:42:07 crc kubenswrapper[4861]: I0315 00:42:07.425708 4861 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b24217a-8fe7-4e28-85d4-8053361942af" path="/var/lib/kubelet/pods/1b24217a-8fe7-4e28-85d4-8053361942af/volumes" Mar 15 00:42:54 crc kubenswrapper[4861]: I0315 00:42:54.704111 4861 scope.go:117] "RemoveContainer" containerID="278a53a72a4bfee42403bf71d89af2f66b89367c0163c28a98249cc164d64705" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.148035 4861 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29558924-fznqv"] Mar 15 00:44:00 crc kubenswrapper[4861]: E0315 00:44:00.148878 4861 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71643876-7f20-406a-8b63-d2e665cd166c" containerName="oc" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.148895 4861 state_mem.go:107] "Deleted CPUSet assignment" podUID="71643876-7f20-406a-8b63-d2e665cd166c" containerName="oc" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.149061 4861 memory_manager.go:354] "RemoveStaleState removing state" podUID="71643876-7f20-406a-8b63-d2e665cd166c" containerName="oc" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.149549 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558924-fznqv" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.152143 4861 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-hg5jl" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.153095 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.157068 4861 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.166748 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558924-fznqv"] Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.312613 4861 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64wbp\" (UniqueName: \"kubernetes.io/projected/7ff706d9-f314-4462-a8d1-2cfde3a0c745-kube-api-access-64wbp\") pod \"auto-csr-approver-29558924-fznqv\" (UID: \"7ff706d9-f314-4462-a8d1-2cfde3a0c745\") " pod="openshift-infra/auto-csr-approver-29558924-fznqv" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.414185 4861 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64wbp\" (UniqueName: \"kubernetes.io/projected/7ff706d9-f314-4462-a8d1-2cfde3a0c745-kube-api-access-64wbp\") pod \"auto-csr-approver-29558924-fznqv\" (UID: \"7ff706d9-f314-4462-a8d1-2cfde3a0c745\") " pod="openshift-infra/auto-csr-approver-29558924-fznqv" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.438434 4861 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64wbp\" (UniqueName: \"kubernetes.io/projected/7ff706d9-f314-4462-a8d1-2cfde3a0c745-kube-api-access-64wbp\") pod \"auto-csr-approver-29558924-fznqv\" (UID: \"7ff706d9-f314-4462-a8d1-2cfde3a0c745\") " pod="openshift-infra/auto-csr-approver-29558924-fznqv" Mar 15 00:44:00 crc kubenswrapper[4861]: I0315 00:44:00.471271 4861 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558924-fznqv" Mar 15 00:44:01 crc kubenswrapper[4861]: I0315 00:44:01.009781 4861 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29558924-fznqv"] Mar 15 00:44:01 crc kubenswrapper[4861]: I0315 00:44:01.026785 4861 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 15 00:44:01 crc kubenswrapper[4861]: I0315 00:44:01.279189 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558924-fznqv" event={"ID":"7ff706d9-f314-4462-a8d1-2cfde3a0c745","Type":"ContainerStarted","Data":"26fc9a6364992a22188aa33dd4d3db51846be179a20c4aeab120bf937b0e9590"} Mar 15 00:44:02 crc kubenswrapper[4861]: I0315 00:44:02.282245 4861 patch_prober.go:28] interesting pod/machine-config-daemon-r8m45 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 15 00:44:02 crc kubenswrapper[4861]: I0315 00:44:02.283106 4861 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8m45" podUID="5442c197-f16b-462d-b7ef-f509c8b1e858" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 15 00:44:02 crc kubenswrapper[4861]: I0315 00:44:02.296965 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558924-fznqv" event={"ID":"7ff706d9-f314-4462-a8d1-2cfde3a0c745","Type":"ContainerStarted","Data":"c0d147153f945f6a91b1aff6aaa0790e2f50a30aaa79b4cd87bfdf209b3f9576"} Mar 15 00:44:02 crc kubenswrapper[4861]: I0315 00:44:02.328357 4861 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29558924-fznqv" podStartSLOduration=1.41171177 podStartE2EDuration="2.328330604s" podCreationTimestamp="2026-03-15 00:44:00 +0000 UTC" firstStartedPulling="2026-03-15 00:44:01.026461058 +0000 UTC m=+2234.338658601" lastFinishedPulling="2026-03-15 00:44:01.943079902 +0000 UTC m=+2235.255277435" observedRunningTime="2026-03-15 00:44:02.315586596 +0000 UTC m=+2235.627784139" watchObservedRunningTime="2026-03-15 00:44:02.328330604 +0000 UTC m=+2235.640528177" Mar 15 00:44:03 crc kubenswrapper[4861]: I0315 00:44:03.308800 4861 generic.go:334] "Generic (PLEG): container finished" podID="7ff706d9-f314-4462-a8d1-2cfde3a0c745" containerID="c0d147153f945f6a91b1aff6aaa0790e2f50a30aaa79b4cd87bfdf209b3f9576" exitCode=0 Mar 15 00:44:03 crc kubenswrapper[4861]: I0315 00:44:03.308865 4861 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29558924-fznqv" event={"ID":"7ff706d9-f314-4462-a8d1-2cfde3a0c745","Type":"ContainerDied","Data":"c0d147153f945f6a91b1aff6aaa0790e2f50a30aaa79b4cd87bfdf209b3f9576"} Mar 15 00:44:04 crc kubenswrapper[4861]: I0315 00:44:04.648690 4861 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29558924-fznqv" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515155400336024447 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015155400336017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015155373453016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015155373454015471 5ustar corecore